Big Data Analytics examines massive amounts of data to uncover hidden patterns, correlations, and other insights. Organizations analyze their data to make smarter business moves, introduce operational efficiencies, yield higher profits, and increase customer satisfaction.

Here are some typical questions you need to ask regarding your big data analytics storage solution:

  • Can I consolidate and synchronize data scattered across many silos?
  • How can I reduce dependency on legacy systems for data processing?
  • Can I ensure data security at all levels?
  • How can I reduce costs for big data that can quickly scale?
  • How do I process data in the cloud and get results on-premises?

Get started now with

Cloud Volumes Service Cloud Volumes ONTAP

Key Considerations

Do you have multiple data repositories in a hybrid or multicloud environment?

Data Consolidation and Cloud Onboarding

A single management view with sync and transfer tools allow data to be easily accessed and migrated to a single cloud environment.

Is minimizing storage costs a key consideration?

Storage Efficiencies

Leverage efficient data snapshots, thin provisioning, data compression, deduplication, data tiering and cloning that can reduce storage expenses by as much as 70%.

How can you guarantee high performance for your big data analytics projects?

Performance and Guaranteed SLAs

When it comes to analyzing data, performance at the highest levels and guaranteed SLAs ensure you get the results you need, when you need them.

Are you looking to speed up dev/test processes for large analytics projects?

Efficient Cloning

NetApp’s cloning technology creates writable volumes from snapshots in zero time helping you streamline and automate fast provisioning of virtual environments.

Choose the offering that best suits your Analytics workloads


Cloud Volumes Service

Fully managed cloud storage service delivered by NetApp

Read More

Cloud Volumes ONTAP

A customer-managed solution providing full ONTAP capabilities

Read More

logo-service-1 Cloud Volumes Service

NetApp’s fully-managed Cloud Volumes Service provides enterprise-grade data management and data protection capabilities with highly-available and highly performant storage for big data analytics workloads.

3 Steps for a Successful POC

  • Sign up for the Cloud Volumes Service trial period
    • Sign up for the Cloud Volumes Service trial period on AWS or register for our Preview Program on Azure or GCP.
    • Enter your information and NetApp will evaluate your requirements and provide you with access.
    • Spin up your first 100 TB of storage in the cloud.
  • Create your first 100 TB in the cloud
    • Log into your account linked to AWS, Azure, or GCP.
    • Name the volume you want to create.
    • Decide which protocol you want to expose.
  • Test how data cloning improves dev/test automation processes, agility and TTM

    Enhance your DevOps environments creation speed with cloning

    • Select a large volume and clone it multiple times.
    • Verify that the volume clones are available for use almost instantly.


  • ic-spin Spin up 100 TB of storage in just seconds
  • ic-utilize Utilize the Cloud Sync service to replicate data to the cloud quickly and efficiently (“lift & shift” data)
  • Use cloning technology to create thin, RW data copies without performance or capacity penalties Use cloning technology

    This will create thin, RW data copies without performance penalties. Such copies can be used as testing environments without compromising production volumes.

  • ic-api Use API calls to automate your allocations

logo-ontap-1 Cloud Volumes ONTAP

Cloud Volumes ONTAP provides enterprise-grade data management you control, with data protection capabilities, highly-available storage, and storage efficiencies in the cloud of your choice.

4 Steps for a Successful POC

  • Test how data cloning improves dev/test processes automation, agility and TTM
    • Select a large volume and clone it multiple times.
    • Verify that the volume clones are available for use almost instantly.
    • Check that the used capacity remains the same, with no increase in storage footprint.
    • Write to the cloned volume and see that the changes are written and that the capacity increased by the changed delta only.
  • Test Cloud Volumes ONTAP HA resiliency
    • Create your Cloud Volumes ONTAP HA environment.
    • Copy data to a Cloud Volumes ONTAP storage volume.
    • Shut down the primary node and verify that the data you copied is accessible.
    • Restart the primary node to resume high availability.
  • Test capacity reductions with storage efficiencies

    Watch your storage footprint reduce and benefit from an average of 50% cost savings.

    • Thin provisioning: Provision a volume which is larger than the aggregate. Despite the overprovisioning, you will not see an increase in the storage footprint.
    • Deduplication: Copy a file multiple times and you’ll see that the used capacity remains the same, with no increase in storage footprint with repeating blocks.
    • Compression: Save a large text file and note how it is compressed substantially, reducing the overall storage footprint.
    • Tiering: Check the volume capacity and see how “cold” data is located on object storage.
  • Evaluate data mobility capabilities with the the Cloud Manager

    See how IT / DevOps teams can manage, monitor, automate, and orchestrate processes:

    • Check the drag-and-drop operation to provision, discover, replicate and sync between on-premises or multicloud environments.
    • Define snapshots policies and sync schedules per storage environment.
    • Examine the history log and actions status in the Timeline view. 
    • Check in the interactive API (Swagger) screens how it enables automation and orchestration.
    • Check multi-tenancy and access control options in the Tenant and User management screens.


  • ew-tip-7 (2) Create your HA environment with the Multiple Availability Zone deployment model to protect against Availability Zone failure
  • Select the Automatic Capacity Management feature Select the Automatic Capacity Management feature

    This will increase the aggregate size automatically when required, and will delete empty aggregates to save storage costs. You can control the thresholds!

  • Leverage the “Schedule Downtime” option Leverage the “Schedule Downtime” option to turn on and off your environment in order to save compute and license costs
  • Tier your cold data, DR environment, and snapshots to Azure Blob Storage or Amazon S3 Tier your DR environment, infrequently-used data, and snapshots to Azure Blob Storage or Amazon S3

    This will reduce storage costs for infrequently-used data to as low as $0.03 per GB per month

Image Title