🔒 Guided

Pre-launch preview. Authorised access only.

Incorrect code

Guided by A Guide to Cloud
Explore AB-900 AI-901
Guided DP-750 Domain 4
Domain 4 — Module 3 of 8 38%
23 of 28 overall

DP-750 Study Guide

Domain 1: Set Up and Configure an Azure Databricks Environment

  • Azure Databricks: Your Lakehouse Platform Free
  • Choosing the Right Compute Free
  • Configuring Compute for Performance Free
  • Unity Catalog: The Three-Level Namespace Free
  • Tables, Views & External Catalogs Free

Domain 2: Secure and Govern Unity Catalog Objects

  • Securing Unity Catalog: Who Gets What
  • Secrets & Authentication
  • Data Discovery & Attribute-Based Access
  • Row Filters, Column Masks & Retention
  • Lineage, Audit Logs & Delta Sharing

Domain 3: Prepare and Process Data

  • Data Modeling: Ingestion Design Free
  • SCD, Granularity & Temporal Tables
  • Partitioning, Clustering & Table Optimization
  • Ingesting Data: Lakeflow Connect & Notebooks
  • Ingesting Data: SQL Methods & CDC
  • Streaming Ingestion: Structured Streaming & Event Hubs
  • Auto Loader & Declarative Pipelines
  • Cleansing & Profiling Data Free
  • Transforming & Loading Data
  • Data Quality & Schema Enforcement

Domain 4: Deploy and Maintain Data Pipelines and Workloads

  • Building Data Pipelines Free
  • Lakeflow Jobs: Create & Configure
  • Lakeflow Jobs: Schedule, Alerts & Recovery
  • Git & Version Control
  • Testing & Databricks Asset Bundles
  • Monitoring Clusters & Troubleshooting
  • Spark Performance: DAG & Query Profile
  • Optimizing Delta Tables & Azure Monitor

DP-750 Study Guide

Domain 1: Set Up and Configure an Azure Databricks Environment

  • Azure Databricks: Your Lakehouse Platform Free
  • Choosing the Right Compute Free
  • Configuring Compute for Performance Free
  • Unity Catalog: The Three-Level Namespace Free
  • Tables, Views & External Catalogs Free

Domain 2: Secure and Govern Unity Catalog Objects

  • Securing Unity Catalog: Who Gets What
  • Secrets & Authentication
  • Data Discovery & Attribute-Based Access
  • Row Filters, Column Masks & Retention
  • Lineage, Audit Logs & Delta Sharing

Domain 3: Prepare and Process Data

  • Data Modeling: Ingestion Design Free
  • SCD, Granularity & Temporal Tables
  • Partitioning, Clustering & Table Optimization
  • Ingesting Data: Lakeflow Connect & Notebooks
  • Ingesting Data: SQL Methods & CDC
  • Streaming Ingestion: Structured Streaming & Event Hubs
  • Auto Loader & Declarative Pipelines
  • Cleansing & Profiling Data Free
  • Transforming & Loading Data
  • Data Quality & Schema Enforcement

Domain 4: Deploy and Maintain Data Pipelines and Workloads

  • Building Data Pipelines Free
  • Lakeflow Jobs: Create & Configure
  • Lakeflow Jobs: Schedule, Alerts & Recovery
  • Git & Version Control
  • Testing & Databricks Asset Bundles
  • Monitoring Clusters & Troubleshooting
  • Spark Performance: DAG & Query Profile
  • Optimizing Delta Tables & Azure Monitor
Domain 4: Deploy and Maintain Data Pipelines and Workloads Premium ⏱ ~10 min read

Lakeflow Jobs: Schedule, Alerts & Recovery

Schedule jobs with cron expressions, configure failure alerts, and set up automatic restarts — making your pipelines self-healing.

Scheduling jobs

☕ Simple explanation

Scheduling is setting an alarm clock for your pipeline.

You tell the job “run every night at 3 AM” or “run every 2 hours on weekdays.” Databricks uses cron expressions — a standard way to describe schedules.

Job scheduling uses cron expressions (minute, hour, day-of-month, month, day-of-week) with timezone support. Jobs can also be configured with concurrency limits (max simultaneous runs) and retry policies for failed tasks.

Common cron patterns

ScheduleCron ExpressionNotes
Every day at 3 AM0 3 * * *Standard nightly ETL
Every 2 hours0 */2 * * *Frequent refresh
Weekdays at 8 AM0 8 * * 1-5Business-hours only
First of every month0 0 1 * *Monthly aggregation

Timezone: Always set the timezone explicitly. “3 AM” means different things in UTC vs NZST.

Alerts

Configure alerts to notify your team when jobs fail or take too long:

Alert TypeWhen It Fires
On failureAny task in the job fails
On successJob completes successfully
On duration exceededJob runs longer than a threshold
On startJob begins execution

Alerts can notify via: email, Slack, PagerDuty, webhooks, or Microsoft Teams.

Ravi configures failure alerts at DataPulse so the on-call engineer gets a Slack message within seconds of any pipeline failure.

Automatic restarts and retries

Task-level retries

Configure retries for individual tasks that may fail due to transient issues:

SettingPurpose
Max retriesNumber of retry attempts (e.g., 3)
Min retry intervalWait time between retries (e.g., 30 seconds)
Retry on timeoutWhether to retry when a task times out

Pipeline automatic restarts

For Declarative Pipelines, configure continuous mode with automatic restart:

  • Pipeline runs continuously
  • If it fails, it automatically restarts after a configurable delay
  • Ideal for streaming pipelines that must stay running

Job-level recovery

FeatureWhat It Does
Repair runRe-run only the failed tasks (not the entire job)
Max concurrent runsPrevent overlapping runs (default: 1)
TimeoutKill the job if it exceeds a time limit
Question

What are the four alert types for Lakeflow Jobs?

Click or press Enter to reveal answer

Answer

On failure, on success, on duration exceeded, and on start. Alerts can notify via email, Slack, PagerDuty, webhooks, or Teams.

Click to flip back

Question

What is a repair run in Lakeflow Jobs?

Click or press Enter to reveal answer

Answer

A repair run re-executes only the failed tasks (and their downstream dependents) rather than restarting the entire job. Saves time and compute when only one task in a multi-task job failed.

Click to flip back

Question

How do you configure automatic restarts for streaming pipelines?

Click or press Enter to reveal answer

Answer

Set the Declarative Pipeline to continuous mode with automatic restart. If it fails, it restarts after a configurable delay. Essential for streaming pipelines that must stay running.

Click to flip back

🎬 Video coming soon

Knowledge check

Knowledge Check

Ravi's nightly ETL job at DataPulse occasionally fails on the ingestion task due to a flaky source API. The API usually recovers within a minute. What configuration minimises manual intervention?


Next up: Git & Version Control — Git best practices, branching, and pull request workflows in Databricks.

← Previous

Lakeflow Jobs: Create & Configure

Next →

Git & Version Control

Guided

I learn, I simplify, I share.

A Guide to Cloud YouTube Feedback

© 2026 Sutheesh. All rights reserved.

Guided is an independent study resource and is not affiliated with, endorsed by, or officially connected to Microsoft. Microsoft, Azure, and related trademarks are property of Microsoft Corporation. Always verify information against Microsoft Learn.