Skip to main content

Build Data Pipelines with Delta Live Tables

In this course, you’ll learn how to define and schedule data pipelines that incrementally ingest and process data through multiple tables in the lakehouse using Delta Live Tables (DLT) in Spark SQL and Python. The course covers how to get started with DLT, how DLT tracks data dependencies in data pipelines, how to configure and run data pipelines using the Delta Live Tables UI, how to use Python or Spark SQL to define data pipelines that ingest and process data through multiple tables in the lakehouse using Auto Loader and DLT, how to use APPLY CHANGES INTO syntax to process Change Data Capture feeds, and how to review event logs and data artifacts created by pipelines and troubleshoot DLT syntax.


Languages Available: English | 日本語 | Português BR | 한국어

Skill Level
Associate
Duration
4h
Prerequisites
  • Beginner-level familiarity with basic cloud concepts (virtual machines, object storage, identity management)

  • Ability to perform basic code development tasks (create compute, run code in notebooks, use basic notebook operations, import repos from git, etc)

  • Intermediate familiarity with basic SQL concepts (CREATE, SELECT, INSERT, UPDATE, DELETE, WHILE, GROUP BY, JOIN, etc.)

Outline

    • The Medallion Architecture
    • Introduction to Delta Live Tables
    • Using the Delta Live Tables UI - PART 1 - Orders
    • Using the Delta Live Tables UI - PART 2 - Customers
    • Using the Delta Live Tables UI - PART 3 - Lab - Status
    • SQL pipelines
    • Python pipelines
    • Delta Live Tables Running Modes
    • Pipeline Results
    • Pipeline Event Logs

Upcoming Public Classes

Date
Time
Language
Price
Jan 06
01 PM - 05 PM (America/New_York)
English
$750.00
Jan 08
01 PM - 05 PM (Asia/Kolkata)
English
$750.00
Jan 10
08 AM - 12 PM (Europe/London)
English
$750.00
Feb 03
01 PM - 05 PM (Europe/London)
English
$750.00
Feb 05
09 AM - 01 PM (America/New_York)
English
$750.00
Feb 07
09 AM - 01 PM (Asia/Kolkata)
English
$750.00
Mar 03
01 PM - 05 PM (Asia/Kolkata)
English
$750.00
Mar 05
01 PM - 05 PM (America/New_York)
English
$750.00
Mar 07
09 AM - 01 PM (Europe/London)
English
$750.00

Public Class Registration

If your company has purchased success credits or has a learning subscription, please fill out the Training Request form. Otherwise, you can register below.

Private Class Request

If your company is interested in private training, please submit a request.

See all our registration options

Registration options

Databricks has a delivery method for wherever you are on your learning journey

Runtime

Self-Paced

Custom-fit learning paths for data, analytics, and AI roles and career paths through on-demand videos

Register now

Instructors

Instructor-Led

Public and private courses taught by expert instructors across half-day to two-day courses

Register now

Learning

Blended Learning

Self-paced and weekly instructor-led sessions for every style of learner to optimize course completion and knowledge retention. Go to Subscriptions Catalog tab to purchase

Purchase now

Scale

Skills@Scale

Comprehensive training offering for large scale customers that includes learning elements for every style of learning. Inquire with your account executive for details

Upcoming Public Classes

Generative AI Engineer

Generative AI Engineering with Databricks

This course is aimed at data scientists, machine learning engineers, and other data practitioners who want to build generative AI applications using the latest and most popular frameworks and Databricks capabilities. 

Below, we describe each of the four, four-hour modules included in this course.

Generative AI Solution Development: This is your introduction to contextual generative AI solutions using the retrieval-augmented generation (RAG) method. First, you’ll be introduced to RAG architecture and the significance of contextual information using Mosaic AI Playground. Next, we’ll show you how to prepare data for generative AI solutions and connect this process with building a RAG architecture. Finally, you’ll explore concepts related to context embedding, vectors, vector databases, and the utilization of Mosaic AI Vector Search.

Generative AI Application Development: Ready for information and practical experience in building advanced LLM applications using multi-stage reasoning LLM chains and agents? In this module, you'll first learn how to decompose a problem into its components and select the most suitable model for each step to enhance business use cases. Following this, we’ll show you how to construct a multi-stage reasoning chain utilizing LangChain and HuggingFace transformers. Finally, you’ll be introduced to agents and will design an autonomous agent using generative models on Databricks.

Generative AI Application Evaluation and Governance: This is your introduction to evaluating and governing generative AI systems. First, you’ll explore the meaning behind and motivation for building evaluation and governance/security systems. Next, we’ll connect evaluation and governance systems to the Databricks Data Intelligence Platform. Third, we’ll teach you about a variety of evaluation techniques for specific components and types of applications. Finally, the course will conclude with an analysis of evaluating entire AI systems with respect to performance and cost.

Generative AI Application Deployment and Monitoring: Ready to learn how to deploy, operationalize, and monitor generative deploying, operationalizing, and monitoring generative AI applications? This module will help you gain skills in the deployment of generative AI applications using tools like Model Serving. We’ll also cover how to operationalize generative AI applications following best practices and recommended architectures. Finally, we’ll discuss the idea of monitoring generative AI applications and their components using Lakehouse Monitoring.

Paid
16h
Lab
instructor-led
Associate

Questions?

If you have any questions, please refer to our Frequently Asked Questions page.