Skip to main content

LakeFlow Connect

Efficient ingestion connectors for all

Bring your data into the Data Intelligence Platform with high efficiency using native ingestion connectors for analytics and AI. With just a few easy steps, create a pipeline that ingests your data without having to author or maintain complex code. Full integration with the Data Intelligence Platform means you get all the benefits of unified governance and observability out of the box, as well as the ability to orchestrate downstream workflows that transform ingested data into insights.

Insulet logo.

“With the new Salesforce ingestion connector from Databricks, we’ve significantly streamlined our data integration process by eliminating fragile and problematic middleware. This improvement allows Databricks SQL to directly analyze Salesforce data within Databricks. As a result, our data practitioners can now deliver updated insights in near real-time, reducing latency from days to minutes.”

— Bill Whiteley, Senior Director of AI, Analytics and Advanced Algorithms, Insulet

An image of a Databricks icon representing better price-performance.

Efficient incremental pipelines

Lower your total cost of ownership and accelerate time to value.

Simple setup and maintenance

Simple setup and maintenance

Democratize data access for data engineers, data analysts and data scientists.

Unified orchestration, observability and governance

Unified orchestration, observability and governance

Take advantage of healthy, secured ingestion pipelines that are native to your platform.

Data ingestion with LakeFlow Connect

Data ingestion with LakeFlow Connectors
Ingest from a growing selection of data sources

Ingest from a growing selection of data sources

Getting the most out of your data means bringing it together from every cloud storage, database and business application you manage so it can be transformed for valuable insights and innovative AI. Choose from a wide variety of connectors to popular data sources so you can unlock the power of your data, no matter where it comes from.

Efficient, incremental ingestion

Efficient, incremental ingestion

Smart optimization brings the best results for every use case so you don’t need to spend time optimizing pipelines for different data sources. When incremental reads and writes are utilized, ingestion pipelines avoid unnecessary work and only bring in new data or table updates. An incremental approach for ingestion is faster, scalable and more cost-efficient, while your data remains fresh for downstream consumption.

An image of a Databricks workspace with a CSV file being read using Spark.

Self-service for every practitioner

Accelerate innovation throughout your organization by empowering every team to access the data they need by letting them define their own custom ingestion pipelines. Data engineers, data analysts and data scientists all benefit from a no-code experience that allows them to get started quickly. They can also use a simple API that saves hours of coding. Moreover, it’s easier to manage pipelines — with less time spent moving data and more time using it.

Ingestion is only the beginning

Ingestion is only the beginning

Bringing all your data efficiently into the Data Intelligence Platform is only the first step in extracting value from your data and accelerating innovation. Orchestrate advanced workflows for analytics and AI and utilize incremental transformations downstream. Build and deploy ML and GenAI applications with Mosaic AI or analyze and visualize your data to extract actionable insights with Databricks SQL.

Your data. Your choice.

With Databricks’ open ecosystem of technology partners, you can choose from 500+ additional pre-built connectors to meet any use case for data engineering.

alteryx
Rivery logo
Prophecy
Fivetran logo
Informatica
Qlik
alteryx
Rivery logo
Prophecy
Fivetran logo
Informatica
Qlik

Ready to become a data + AI company?

Take the first steps in your transformation