Skip to main content
Company Blog

Continued Innovation and Expanded Availability for the Next-gen Unified Analytics Engine

Databricks Delta Lake the next generation unified analytics engine, built on top of Apache Spark, and aimed at helping data engineers build robust production data pipelines at scale is continuing to make strides. Already a powerful approach to building data pipelines, new capabilities and performance enhancements make Delta an even more compelling solution:

Fast Parquet import: makes it easy to convert existing Parquet files for use with Delta. The conversion can now be completed much more quickly and without requiring large amount of extra compute and storage resources. Please see the documentation for more details: Azure | AWS.

Example:

Time Travel: allows you to use and revert to earlier versions of your data. This is useful for conducting analyses that depend upon earlier versions of the data or for correcting errors. Please see this recent blog or the documentation for more details: Azure | AWS.

Example:

MERGE enhancements: In order to enable writing change data (inserts, updates, and deletes) from a database to a Delta table, there is now support for multiple MATCHED clauses, additional conditions in MATCHED and NOT MATCHED clauses, and a DELETE action. There is also support for * in UPDATE and INSERT actions to automatically fill in column names (similar to * in SELECT). This makes it easier to write MERGE queries for tables with a very large number of columns. Please see the documentation for more details: Azure | AWS.

Examples:

MERGE <b>INTO</b> [db_name.]target_table [<b>AS</b> target_alias]
<b>USING</b> [db_name.]source_table [<b>AS</b> source_alias]
<b>ON</b> <merge_condition>
[ <b>WHEN</b> MATCHED [ <b>AND</b> <condition> ] <b>THEN</b> <matched_action> ]
[ <b>WHEN</b> MATCHED [ <b>AND</b> <condition> ] <b>THEN</b> <matched_action> ]
[ <b>WHEN</b> <b>NOT</b> MATCHED [ <b>AND</b> <condition> ]  <b>THEN</b> <not_matched_action> ]

Where,

<matched_action>  =
 <b>DELETE</b>  |
 <b>UPDATE</b> <b>SET</b> *  |
 <b>UPDATE</b> <b>SET</b> column1 = value1 [, column2 = value2 ...]
<not_matched_action>  =
 <b>INSERT</b> *  |
 <b>INSERT</b> (column1 [, column2 ...]) <b>VALUES</b> (value1 [, value2 ...])</not_matched_action></matched_action></not_matched_action></condition></matched_action></condition></matched_action></condition></merge_condition>

Expanded Availability

In response to rising customer interest we have decided to make Databricks Delta more broadly available. Azure Databricks customers now have access to Delta capabilities for Data Engineering and Data Analytics from both the Azure Databricks Standard and the Azure Databricks Premium SKUs.  Similarly, customers using Databricks on AWS now have access to Delta from both Data Engineering and Data Analytics offerings.

CloudOfferingDelta Included?
Microsoft AzureAzure Databricks Standard Data Engineering      ✓
Azure Databricks Standard Data Analytics      ✓
Azure Databricks Premium Data Engineering      ✓
Azure Databricks Premium Data Analytics      ✓
AWSDatabricks Basic      ✘
Databricks Data Engineering      ✓
Databricks Data Analytics      ✓

Easy to Adopt: Check Out Delta Today

Porting existing Spark code for using Delta is as simple as changing

“CREATE TABLE … USING parquet” to
“CREATE TABLE … USING delta”

or changing

“dataframe.write.format(“parquet“).load(“/data/events“)” to
“dataframe.write.format(“delta“).load(“/data/events“)”

You can explore Delta today using:

  • Databricks Delta Quickstart - for an introduction to Databricks Delta Azure | AWS.
  • Optimizing Performance and Cost - for a discussion of features such as compaction, z-ordering and data skipping Azure | AWS.

Both of these contain notebooks in Python, Scala and SQL that you can use to try Delta.

If you are not already using Databricks, you can try Databricks Delta for free by signing up for the Databricks trial Azure | AWS.

Making Data Engineering Easier

Data engineering is critical to successful analytics and customers can use Delta in various way to improve their data pipelines. We have summarized some of these use cases in the below set of blogs:

You can learn more about Delta from the Databricks Delta documentation Azure | AWS.

Interested in the open source Delta Lake?
Visit the Delta Lake online hub to learn more, download the latest code and join the Delta Lake community.