Skip to main content

Introducing Databricks LakeFlow: A unified, intelligent solution for data engineering

Ingest data from databases, enterprise apps and cloud sources, transform it in batch and real-time streaming, and confidently deploy and operate in production.
Share this post

Today, we are excited to announce Databricks LakeFlow, a new solution that contains everything you need to build and operate production data pipelines. It includes new native, highly scalable connectors for databases including MySQL, Postgres, SQL Server and Oracle and enterprise applications like Salesforce, Microsoft Dynamics, NetSuite, Workday, ServiceNow and Google Analytics. Users can transform data in batch and streaming using standard SQL and Python. We are also announcing Real Time Mode for Apache Spark, allowing stream processing at orders of magnitude faster latencies than microbatch. Finally, you can orchestrate and monitor workflows and deploy to production using CI/CD. Databricks LakeFlow is native to the Data Intelligence Platform, providing serverless compute and unified governance with Unity Catalog.

LakeFlow
LakeFlow is the one unified data engineering solution for ingestion, transformation and orchestration

In this blog post we discuss the reasons why we believe LakeFlow will help data teams meet the growing demand of reliable data and AI as well as LakeFlow’s key capabilities integrated into a single product experience.

Challenges in building and operating reliable data pipelines

Data engineering - collecting and preparing fresh, high-quality and reliable data - is a necessary ingredient for democratizing data and AI in your business. Yet achieving this remains full of complexity and requires stitching together many different tools.

 

First, data teams need to ingest data from multiple systems each with their own formats and access methods. This requires building and maintaining in-house connectors for databases and enterprise applications. Just keeping up with enterprise applications’ API changes can be a full-time job for an entire data team. Data then needs to be prepared in both batch and streaming, which requires writing and maintaining complex logic for triggering and incremental processing. When latency spikes or a failure occurs, it means getting paged, a set of unhappy data consumers and even disruptions to the business that affect the bottom line. Finally, data teams need to deploy these pipelines using CI/CD and monitor the quality and lineage of data assets. This normally requires deploying, learning and managing another entirely new tool like Prometheus or Grafana.

 

This is why we decided to build LakeFlow, a unified solution for data ingestion, transformation, and orchestration powered by data intelligence. Its three key components are: LakeFlow Connect, LakeFlow Pipelines and LakeFlow Jobs.

LakeFlow Connect: Simple and scalable data ingestion 

LakeFlow Connect provides point-and-click data ingestion from databases such as MySQL, Postgres, SQL Server and Oracle and enterprise applications like Salesforce, Microsoft Dynamics, NetSuite, Workday, ServiceNow and Google Analytics. LakeFlow Connect can also ingest unstructured data such as PDFs and Excel spreadsheets from sources like SharePoint.

 

It extends our popular native connectors for cloud storage (e.g. S3, ADLS Gen2 and GCS) and queues (e.g. Kafka, Kinesis, Event Hub and Pub/Sub connectors), and partner solutions such as Fivetran, Qlik and Informatica.

LakeFlow Connector
Setup an ingestion pipeline in a few easy steps with LakeFlow Connect

We are particularly excited about database connectors, which are powered by our acquisition of Arcion. An incredible amount of valuable data is locked away in operational databases. Instead of naive approaches to load this data, which hit operational and scaling issues, LakeFlows uses change data capture (CDC) technology to make it simple, reliable and operationally efficient to bring this data to your lakehouse.

 

Databricks customers who are using LakeFlow Connect find that a simple ingestion solution improves productivity and lets them move faster from data to insights. Insulet, a manufacturer of a wearable insulin management system, the Omnipod, uses the Salesforce ingestion connector to ingest data related to customer feedback into their data solution which is built on Databricks. This data is made available for analysis through Databricks SQL to gain insights regarding quality issues and track customer complaints. The team found significant value in using the new capabilities of LakeFlow Connect.

"With the new Salesforce ingestion connector from Databricks, we've significantly streamlined our data integration process by eliminating fragile and problematic middleware. This improvement allows Databricks SQL to directly analyze Salesforce data within Databricks. As a result, our data practitioners can now deliver updated insights in near-real time, reducing latency from days to minutes."
— Bill Whiteley, Senior Director of AI,  Analytics, and Advanced Algorithms, Insulet

LakeFlow Pipelines: Efficient declarative data pipelines

LakeFlow Pipelines lower the complexity of building and managing efficient batch and streaming data pipelines. Built on the declarative Delta Live Tables framework, they free you up to write business logic in SQL and Python while Databricks automates data orchestration, incremental processing and compute infrastructure autoscaling on your behalf. Moreover, LakeFlow Pipelines offers built-in data quality monitoring and its Real Time Mode lets you enable consistently low-latency delivery of time-sensitive datasets without any code changes.

LakeFlow Pipelines simplifies data pipeline automation
LakeFlow Pipelines simplifies data pipeline automation

LakeFlow Jobs: Reliable orchestration for every workload

LakeFlow Jobs reliably orchestrates and monitors production workloads. Built on the advanced capabilities of Databricks Workflows, it orchestrates any workload, including ingestion, pipelines, notebooks, SQL queries, machine learning training, model deployment and inference. Data teams can also leverage triggers, branching and looping to meet complex data delivery use cases.

 

LakeFlow Jobs also automates and simplifies the process of understanding and tracking data health and delivery. It takes a data-first view of health, giving data teams full lineage including relationships between ingestion, transformations, tables and dashboards. Additionally, it tracks data freshness and quality, allowing data teams to add monitors via Lakehouse Monitoring with the click of a button.

Built on the Data Intelligence Platform

Databricks LakeFlow is natively integrated with our Data Intelligence Platform, which brings these capabilities:

  • Data intelligence: AI-powered intelligence is not just a feature of LakeFlow, it is a foundational capability that touches every aspect of the product. Databricks Assistant powers the discovery, authoring and monitoring of data pipelines, so you can spend more time building reliable data.
  • Unified governance: LakeFlow is also deeply integrated with Unity Catalog, which powers lineage and data quality. 
  • Serverless compute: Build and orchestrate pipelines at scale and help your team focus on work without having to worry about infrastructure.

The future of data engineering is simple, unified and intelligent

We believe that LakeFlow will enable our customers to deliver fresher, more complete and higher-quality data to their businesses. LakeFlow will enter preview soon starting with LakeFlow Connect. If you would like to request access, sign up here. Over the coming months, look for more LakeFlow announcements as additional capabilities become available.

Try Databricks for free

Related posts

See all Platform Blog posts