Design and implement production ready Lakehouse architectures using Delta Lake and Databricks. By the end of this course, you will be able to build multi layer Medallion pipelines including Bronze, Silver, and Gold layers, manage ACID transactions, enforce and evolve schemas, implement Change Data Capture, and optimize Delta tables for performance using data skipping, compaction, and Liquid Clustering. You will also learn to unify batch and streaming workloads while ensuring reliability, scalability, and recoverability in enterprise environments.

Lakehouse Architecture and Delta Lake with Databricks

Recommended experience
What you'll learn
Design and implement Lakehouse architectures using Databricks and Delta Lake to replace legacy data platforms
Build end-to-end data pipelines using Medallion Architecture (Bronze, Silver, Gold) with incremental processing and Change Data Capture
Apply Delta Lake performance optimization techniques—including data skipping, file compaction, and Liquid Clustering—to support BI and ML workloads
Manage production-grade data reliability through ACID transactions, time travel, schema enforcement, and concurrency control
Details to know

Add to your LinkedIn profile
February 2026
See how employees at top companies are mastering in-demand skills

There are 4 modules in this course
This module introduces the evolution of modern data platforms, from traditional warehouses and data lakes to the unified Lakehouse architecture. Learners explore foundational concepts of Databricks, Apache Spark, and Delta Lake that enable scalable, reliable, and governed data processing.
What's included
15 videos5 readings4 assignments
This module focuses on the core operational capabilities of Delta Lake, including storage architecture, metadata management, transactional processing, and schema control. Learners gain hands-on experience with CRUD operations, incremental data pipelines, time travel, and streaming to build reliable, production-ready data workflows.
What's included
12 videos4 readings4 assignments
This module focuses on designing scalable Lakehouse architectures using Medallion patterns and optimizing Delta Lake for performance and cost efficiency. Learners build multi-layer data pipelines and apply advanced optimization techniques to support BI and machine learning workloads.
What's included
13 videos5 readings4 assignments
This module focuses on designing scalable Lakehouse architectures using Medallion patterns and optimizing Delta Lake for performance and cost efficiency. Learners build multi-layer data pipelines and apply advanced optimization techniques to support BI and machine learning workloads.
What's included
9 videos3 readings4 assignments
Why people choose Coursera for their career

Felipe M.

Jennifer J.

Larry W.

Chaitanya A.

Open new doors with Coursera Plus
Unlimited access to 10,000+ world-class courses, hands-on projects, and job-ready certificate programs - all included in your subscription
Advance your career with an online degree
Earn a degree from world-class universities - 100% online
Join over 3,400 global companies that choose Coursera for Business
Upskill your employees to excel in the digital economy
Frequently asked questions
Delta Lake is an open-source storage layer that brings ACID transactions, schema enforcement, and time travel to data lakes. This course teaches you to use Delta Lake natively on Databricks to build reliable, scalable data pipelines.
No. The course starts from the Databricks environment and workspace setup. Basic SQL and Python knowledge is recommended, but no prior Databricks or Delta Lake experience is required.
A Lakehouse combines the low-cost storage of a data lake with the reliability and performance of a data warehouse. You'll learn how Databricks implements this unified architecture using Delta Lake as its core.
More questions
Financial aid available,
¹ Some assignments in this course are AI-graded. For these assignments, your data will be used in accordance with Coursera's Privacy Notice.


