Lakehouse as Code | 04. Delta Live Tables Data Pipelines
Welcome to the Lakehouse as Code mini-series! In this series, we'll walk you through deploying a complete Databricks lakehouse using infrastructure as code with Laktory. From setting up Unity Catalog to orchestrating data pipelines and configuring your workspace, we’ve got everything covered.
In this fourht part, we focus on configuring a Databricks workspace. You’ll learn how to:
Configure a Laktory data pipeline as a Delta Live Tables
Declare a multi-tables data pipeline
Set up streaming tables with incremental data processing
Develop and debug the pipeline from your IDE
Deploy and execute as Delta Live Tables