Spark Declarative Pipelines (SDP), introduced in Apache Spark 4.1, extends declarative programming from individual queries to entire data pipelines. Instead of manually managing incremental processing, dependency ordering, data quality checks, backfills, and orchestration via external tools like Airflow, engineers declare what

6m read time From databricks.com
Post cover image

Sort: