Databricks on AWS

Databricks on AWS pricing comes in Standard, Standard with security add on, and Enterprise packages.
Each package has different price points for Data Engineering and Data Analytics workloads.
We also offer Basic tier as a low-cost option for Data Engineering workloads.

Databricks Data
Engineering
Description
Automated workloads to run fast
and robust jobs via API or UI.
Pay based on the workload you run
Start your trial
$0.20per Databricks Unit (DBU) + AWS cost
Apply for free AWS credit
Apache Spark
Clusters for running production jobs
Alerting and monitoring with retries
Production streaming with monitoring
Scheduler for running libraries and notebooks
Notebook workflows
Optimized automatic scaling of compute
Automatic scaling of local storage
SSD Caching
Optimized storage access layer (S3, Redshift)
Faster data reading with columnar formats
Strong reliability during data writes to storage
Kinesis and S3-SQS Support for Structured Streaming
Spark / Horovod Integration
XGBoost Support
TensorFlow & Keras Support
Persistent clusters for analytics
Multi-user sharing and collaboration per cluster
SQL, Python, R and Scala notebooks
One-click visualization
Interactive dashboards
Collaboration
Revision history
Version Control System integrations (Github, Bitbucket)
RStudio Integration NEW!
Optional add-on package
+ $0.15per Databricks Unit (DBU)
Role-based access control for notebooks, clusters, jobs, and structured data
Single sign-on (SSO) with SAML 2.0 support
JDBC/ODBC authentication
Single-tenant and compliance
Custom pricing
All standard plan features
All Operational Security Package features
Single-tenant deployment
Audit logs
Customer-provided KMS key for encrypting data
No public IPs for worker nodes
Restricted network access for end users
Customized CIDR range of deployment
HIPAA
AWS GovCloud
Databricks Data
Analytics
Description
Interactive workloads to analyze data
collaboratively with notebooks.
Pay based on the workload you run
Start your trial
$0.40per Databricks Unit (DBU) + AWS cost
Apply for free AWS credit
Apache Spark
Clusters for running production jobs
Alerting and monitoring with retries
Production streaming with monitoring
Scheduler for running libraries and notebooks
Notebook workflows
Optimized automatic scaling of compute
Automatic scaling of local storage
SSD Caching
Optimized storage access layer (S3, Redshift)
Faster data reading with columnar formats
Strong reliability during data writes to storage
Kinesis and S3-SQS Support for Structured Streaming
Spark / Horovod Integration
XGBoost Support
TensorFlow & Keras Support
Persistent clusters for analytics
Multi-user sharing and collaboration per cluster
SQL, Python, R and Scala notebooks
One-click visualization
Interactive dashboards
Collaboration
Revision history
Version Control System integrations (Github, Bitbucket)
RStudio Integration NEW!
Optional add-on package
+ $0.15per Databricks Unit (DBU)
Role-based access control for notebooks, clusters, jobs, and structured data
Single sign-on (SSO) with SAML 2.0 support
JDBC/ODBC authentication
Single-tenant and compliance
Custom pricing
All standard plan features
All Operational Security Package features
Single-tenant deployment
Audit logs
Customer-provided KMS key for encrypting data
No public IPs for worker nodes
Restricted network access for end users
Customized CIDR range of deployment
HIPAA
AWS GovCloud
NEW
Databricks Basic
in private preview
Description
Automated workloads to run jobs at low cost.
Pay based on the workload you run
Start your trial
$0.07per Databricks Unit (DBU) + AWS cost
Apply for free AWS credit
Apache Spark
Clusters for running production jobs
Alerting and monitoring with retries
Production streaming with monitoring
Scheduler for running libraries and notebooks
Notebook workflows
Optimized automatic scaling of compute
Automatic scaling of local storage
SSD Caching
Optimized storage access layer (S3, Redshift)
Faster data reading with columnar formats
Strong reliability during data writes to storage
Kinesis and S3-SQS Support for Structured Streaming
Spark / Horovod Integration
XGBoost Support
TensorFlow & Keras Support
Persistent clusters for analytics
Multi-user sharing and collaboration per cluster
SQL, Python, R and Scala notebooks
One-click visualization
Interactive dashboards
Collaboration
Revision history
Version Control System integrations (Github, Bitbucket)
RStudio Integration NEW!
Optional add-on package
Role-based access control for notebooks, clusters, jobs, and structured data
Single sign-on (SSO) with SAML 2.0 support
JDBC/ODBC authentication
Single-tenant and compliance
All standard plan features
All Operational Security Package features
Single-tenant deployment
Audit logs
Customer-provided KMS key for encrypting data
No public IPs for worker nodes
Restricted network access for end users
Customized CIDR range of deployment
HIPAA
AWS GovCloud
NEW
Databricks Basic
in private preview
Databricks Data
Engineering
Databricks Data
Analytics
Description
Automated workloads to run jobs at low cost.
Automated workloads to run fast
and robust jobs via API or UI.
Interactive workloads to analyze data
collaboratively with notebooks.
Pay based on the workload you run
Start your trial
$0.07per Databricks Unit (DBU) + AWS cost
Apply for free AWS credit
$0.20per Databricks Unit (DBU) + AWS cost
Apply for free AWS credit
$0.40per Databricks Unit (DBU) + AWS cost
Apply for free AWS credit
Apache Spark
Clusters for running production jobs
Alerting and monitoring with retries
Production streaming with monitoring
Scheduler for running libraries and notebooks
Notebook workflows
Optimized automatic scaling of compute
Automatic scaling of local storage
SSD Caching
Optimized storage access layer (S3, Redshift)
Faster data reading with columnar formats
Strong reliability during data writes to storage
Kinesis and S3-SQS Support for Structured Streaming
Spark / Horovod Integration
XGBoost Support
TensorFlow & Keras Support
Persistent clusters for analytics
Multi-user sharing and collaboration per cluster
SQL, Python, R and Scala notebooks
One-click visualization
Interactive dashboards
Collaboration
Revision history
Version Control System integrations (Github, Bitbucket)
RStudio Integration NEW!
Optional add-on package
+ $0.15per Databricks Unit (DBU)
Role-based access control for notebooks, clusters, jobs, and structured data
Single sign-on (SSO) with SAML 2.0 support
JDBC/ODBC authentication
Single-tenant and compliance
Custom pricing
All standard plan features
All Operational Security Package features
Single-tenant deployment
Audit logs
Customer-provided KMS key for encrypting data
No public IPs for worker nodes
Restricted network access for end users
Customized CIDR range of deployment
HIPAA
AWS GovCloud

Professional Services

Founded by the team who started the Spark research project at UC Berkeley, Databricks is the the best team to help your Spark project.
Contact us

Top Trainers

With the best trainers in the industry, we have taught thousands of classes across hundreds of companies.
Contact us

Best in Class Support

With a large number of Apache Spark committers, Databricks can answer all your Spark questions.
Learn more

Sue runs a daily ETL job that prepares server logs data for analysis and model training. The job runs for 2 hours daily on 10 c3.4xlarge instances.

  • 10 c3.4xlarge nodes
  • × 2 hours
  • × 2 DBUs per hour
  • × $0.20 per automated workload DBU
  • × 30 days
= $240.00 per month
+

Sue’s team also analyzes data (run ad hoc queries) with notebooks for 4 hours on 5 memory-optimized r3.xlarge instance types.

  • 5 r3.xlarge nodes
  • × 4 hours
  • × 1 DBUs per hour
  • × $0.40 per interactive workload DBU
  • × 30 days
= $240.00 per month
+

Sue’s team also uses the Databricks Operational Security add-on.

  • 10 c3.4xlarge nodes
  • × 2 hours
  • × 2 DBUs per hour
  • × 30 days
+
  • 5 r3.xlarge nodes
  • × 4 hours
  • × 1 DBUs per hour
  • × 30 days
$0.15 per DBU
= $270.00 per month
Total cost excluding AWS instance costs:
$750.00 per month

What is a DBU?

A Databricks Unit (“DBU”) is a unit of processing capability per hour, billed on per-second usage. Databricks supports many instance types, see the full list and details here.

I see three different prices for workloads – Data Engineering, Data Analytics and Basic. How can I pick the one that I want?

Our free trial and standard subscription includes both Data Engineering and Data Analytics packages. You don’t need to pick one or the other. Databricks charges you different prices depending on how you use the Apache Spark clusters. You can schedule robust and fast jobs via API or UI at $0.20 per DBU and use persistent clusters for running interactive analytics with notebooks at $0.40 per DBU.

Databricks Basic is currently in Private Preview. Please contact us if you are interested.

How do you define “usage?”

Usage is defined as running Apache Spark clusters, whether they are actively used or not.

What’s the difference between automated Data Engineering and interactive Data Analytics workloads?

An automated workload is defined as job that both starts and terminates the cluster on which it runs. For example, a workload may be triggered by the Databricks Job Scheduler which launches a new Apache Spark cluster solely for the job and automatically terminates the cluster after the job is complete.

An interactive workload is any workload that is not an automated workload, e.g., running a command within Databricks notebooks. These commands run on Apache Spark clusters that may persist until manually terminated. Multiple users can share a cluster for doing interactive analysis in a collaborative way.

Databricks Basic workloads are automated-only workloads. Please contact us for more details on using Databricks Basic.

Databricks Operational Security has an additional charge for DBU usage. Does it apply to all workload types?

Databricks Operational Security applies to both Data Engineering and Data Analytics, but is not available for Databricks Basic. When you choose to deploy Databricks with Databricks Operational Security, the additional charge applies to both Data Engineering and Data Analytics workloads.

How do I add Databricks Operational Security to my subscription?

Please contact us to add Databricks Operational Security to your subscription.

What does the free trial include?

The 14-day free trial gives you access to all features in the Databricks for Data Engineering and Databricks for Data Analytics, except the Databricks Operational Security Package which is an add-on and Databricks Basic, which is in private preview. Contact us if you are interested in trying Databricks Operational Security and / or Databricks Basic.

During trial, AWS will bill you directly for the EC2 instances created in Databricks.

What happens after the free trial?

At the end of the trial, you are automatically subscribed to the Databricks Standard Plan without Databricks Enterprise Security. You can cancel your subscription at any time.

What is Databricks Community Edition?

Databricks Community Edition is a free, limited functionality platform designed for anyone who wants to learn Spark. Sign up here.

How do I trial the Enterprise Plan?

Please contact us to start an Enterprise trial.

Do you provide technical support?

We offer technical support with annual commitments. Contact us to learn more or get started.

How will I be billed?

By default, you will be billed monthly based on per-second usage on your credit card. Contact us for more billing options, such as billing by invoice or an annual plan.