Databricks SQL (DB SQL) allows customers to operate a multi-cloud lakehouse architecture that provides up to 6x better price/performance than traditional cloud data warehouses. Using open source standards to avoid data lock-in, it provides the reliability, quality and performance capabilities that data lakes natively lack.
Reliable, lightning-fast analytics on data lake data
Gain a competitive edge by running SQL queries on your lakehouse with data warehousing performance at data lake economics. DB SQL lets you bring reliability, quality, scale, security and performance to your data lake to support traditional analytics workloads using your most recent and complete data.
Simplified administration and fine-grained governance
Quickly enable data analysts with serverless SQL compute. DB SQL automatically manages instance types, configuration, and provides you with the best price/performance. Granular logging provides visibility into how data is being accessed and queried across your lakehouse so you can maintain data compliance and security, triage errors and troubleshoot execution when needed.
Analytics on all your data with your tools of choice
Connect your preferred BI tools to analyze your most recent and complete data without moving any data to a data warehouse. DB SQL also lets you easily query and transform your data lake data using a built-in SQL editor, build visualizations, and share interactive dashboards that stay up to date.
How does it work?
Up to 6x better price/performance for query execution
Databricks SQL is packed with thousands of optimizations to provide you with the best performance for all query types and real-world applications. This includes Photon — the next-generation query engine — which provides up to 6x better price/performance compared to other cloud data warehouses.
Simplified administration and governance for your lakehouse
Databricks SQL makes it easy to set up and manage SQL compute resources thanks to a central log that records usage across virtual clusters, users and time. This makes it easier to observe workloads across DB SQL, third-party BI tools and any other SQL clients in one place, which in turn helps triage errors and performance issues. Administrators can then drill down into the phases of each query’s execution to troubleshoot problems and support audits.
Connect with your existing tools
Connect your preferred BI tools and benefit from fast performance, low latency and high user concurrency to your data lake data. Setting up reliable connections to your Delta Lake tables is simple, and you can integrate your existing authentication solution. Re-engineered ODBC/JDBC drivers provide lower latency and less overhead to reduce round trips by 0.25 seconds. Data transfer rate is improved 50%, and metadata retrieval operations execute up to 10x faster.
First-class SQL development experience
Databricks SQL allows data analysts to quickly discover and find data sets, write queries in a familiar SQL syntax and easily explore Delta Lake table schemas for ad hoc analysis. Regularly used SQL code can be saved as snippets for quick reuse, and query results can be cached to keep run times short.
Quickly discover and share new insights
Analysts can easily make sense of query results through a wide variety of rich visualizations, and quickly build dashboards with an intuitive drag-and-drop interface. To keep everyone current, dashboards can be shared and configured to automatically refresh, as well as to alert the team to meaningful changes in the data.
Enable your lakehouse
Connect your BI tools to one source of truth for all your data
Collaboratively explore the latest
and freshest data
Now more than ever, organizations need a data strategy that enables speed and agility to be adaptable. As organizations are rapidly moving their data to the cloud, we’re seeing growing interest in doing analytics on the data lake. The introduction of Databricks SQL delivers an entirely new experience for customers to tap into insights from massive volumes of data with the performance, reliability and scale they need. We’re proud to partner with Databricks to bring that opportunity to life.
—Francois Ajenstat, Chief Product Officer, Tableau
Data + AI Summit
Building the Lakehouse at Atlassian
Building omnichannel loyalty and engagement for retailers
Driving Transformation With Scalable, Open Lakehouse Architecture
Data + AI Summit
Powering telemetry analysis at Comcast with Databricks SQL
Data + AI Summit
Delivering insights from 20M+ smart homes with 500M+ devices
“Shell has been undergoing a digital transformation as part of our ambition to deliver more and cleaner energy solutions. As part of this, we have been investing heavily in our data lake architecture. Our ambition has been to enable our data teams to rapidly query our massive datasets in the simplest possible way. The ability to execute rapid queries on petabyte scale datasets using standard BI tools is a game changer for us.
Our co-innovation approach with Databricks has allowed us to influence the product roadmap and we are excited to see this come to market.”
— Dan Jeavons, GM Data Science
“At Atlassian, we need to ensure teams can collaborate well across functions to achieve constantly evolving goals. A simplified lakehouse architecture would empower us to ingest high volumes of user data and run the analytics necessary to better predict customer needs and improve the experience of our customers.
A single, easy-to-use cloud analytics platform allows us to rapidly improve and build new collaboration tools based on actionable insights.”
— Rohan Dhupelia, Data Platform Senior Manager, Atlassian
“At Wejo, we’re collecting data from more than 50 million accessible connected cars to build a better driving experience.
Databricks and a robust lakehouse architecture will allow us to provide automated analytics to our customers, empowering them to glean insights on nearly 5 trillion data points per month, all in a streaming environment from car to marketplace in seconds.”
— Daniel Tibble, Head of Data, wejo
“As a company focused on providing data-driven research to our customers, the massive amount of data in our data lake is our lifeblood. By leveraging Databricks and Delta Lake, we have already been able to democratize data at scale, while lowering the cost of running production workloads by 60%, saving us millions of dollars.
We’re excited to build on this momentum by leveraging the Databricks lakehouse architecture that will further empower everyone across our organization – from research analysts to data scientists – to interchangeably use the same data, helping us to provide innovative insights to our customers faster than ever before.”
— Steve Pulec, Chief Technology Officer, YipitData