🔒 Guided

Pre-launch preview. Authorised access only.

Incorrect code

Guided by A Guide to Cloud
Explore AB-900 AI-901
Guided DP-900 Domain 1
Domain 1 — Module 5 of 7 71%
5 of 27 overall

DP-900 Study Guide

Domain 1: Core Data Concepts

  • Your First Look at Data Free
  • Data File Formats: CSV, JSON, Parquet & More Free
  • Databases: Relational vs Non-Relational Free
  • Transactional Workloads: Keeping Data Consistent Free
  • Analytical Workloads: Finding the Insights Free
  • Data Roles: DBA, Engineer & Analyst Free
  • The Azure Data Landscape Free

Domain 2: Relational Data on Azure

  • Relational Data: Tables, Keys & Relationships
  • Normalization: Why Duplicate Data is Bad
  • SQL Basics: SELECT, INSERT, UPDATE, DELETE
  • Database Objects: Views, Indexes & More
  • Azure SQL: Your Database in the Cloud
  • Open-Source Databases on Azure
  • Choosing the Right Azure Database

Domain 3: Non-Relational Data on Azure

  • Azure Blob Storage: Files in the Cloud
  • Azure Files & Table Storage
  • Azure Cosmos DB: The Global Database
  • Cosmos DB APIs: SQL, MongoDB & More
  • Choosing Non-Relational Storage

Domain 4: Analytics on Azure

  • Data Ingestion & Processing
  • Analytical Data Stores: Data Lakes, Warehouses & Lakehouses
  • Microsoft Fabric & Azure Databricks
  • Batch vs Streaming: Two Speeds of Data
  • Real-Time Analytics on Azure
  • Power BI: See Your Data
  • Data Models in Power BI
  • Choosing the Right Visualization

DP-900 Study Guide

Domain 1: Core Data Concepts

  • Your First Look at Data Free
  • Data File Formats: CSV, JSON, Parquet & More Free
  • Databases: Relational vs Non-Relational Free
  • Transactional Workloads: Keeping Data Consistent Free
  • Analytical Workloads: Finding the Insights Free
  • Data Roles: DBA, Engineer & Analyst Free
  • The Azure Data Landscape Free

Domain 2: Relational Data on Azure

  • Relational Data: Tables, Keys & Relationships
  • Normalization: Why Duplicate Data is Bad
  • SQL Basics: SELECT, INSERT, UPDATE, DELETE
  • Database Objects: Views, Indexes & More
  • Azure SQL: Your Database in the Cloud
  • Open-Source Databases on Azure
  • Choosing the Right Azure Database

Domain 3: Non-Relational Data on Azure

  • Azure Blob Storage: Files in the Cloud
  • Azure Files & Table Storage
  • Azure Cosmos DB: The Global Database
  • Cosmos DB APIs: SQL, MongoDB & More
  • Choosing Non-Relational Storage

Domain 4: Analytics on Azure

  • Data Ingestion & Processing
  • Analytical Data Stores: Data Lakes, Warehouses & Lakehouses
  • Microsoft Fabric & Azure Databricks
  • Batch vs Streaming: Two Speeds of Data
  • Real-Time Analytics on Azure
  • Power BI: See Your Data
  • Data Models in Power BI
  • Choosing the Right Visualization
Domain 1: Core Data Concepts Free ⏱ ~10 min read

Analytical Workloads: Finding the Insights

Transactional systems record what happens. Analytical systems answer why it happened and what to do next. Let's explore the world of OLAP, data warehouses, and reporting.

What is an analytical workload?

☕ Simple explanation

Analytical workloads answer big-picture questions using lots of data.

Imagine Priya at FreshMart. The point-of-sale system records every transaction (that’s transactional). But Priya’s job is different — she asks questions like “Which stores had the highest sales last quarter?” or “Are organic products growing faster than conventional?”

To answer those questions, she doesn’t need individual receipts. She needs summaries, trends, and comparisons across millions of records. That’s an analytical workload.

An analytical workload (OLAP — Online Analytical Processing) processes large volumes of historical data to identify patterns, trends, and insights. Unlike transactional systems that optimise for individual record operations, analytical systems optimise for aggregation queries — SUM, COUNT, AVG, GROUP BY — across millions or billions of records.

Analytical data is typically loaded from transactional sources through ETL/ELT pipelines, stored in data warehouses or data lakes, and consumed through reporting tools like Power BI.

How analytical workloads differ from transactional

FeatureTransactional (OLTP)Analytical (OLAP)
Question type”What just happened?""What happened over time?”
Data volume per querySingle records or small batchesMillions to billions of records
OperationsINSERT, UPDATE, DELETESELECT with aggregations (SUM, AVG, COUNT)
Data freshnessReal-time (current state)Near-real-time to periodic (often loaded in batches)
UsersOperational staff (cashiers, dispatchers)Analysts, managers, executives
Schema designNormalised (reduce duplication)Denormalised (optimise for reading)
Azure servicesAzure SQL DatabaseMicrosoft Fabric, Azure Databricks, Power BI

The analytical pipeline

Data doesn’t start in an analytical system — it flows there from transactional sources through a pipeline:

  1. Source systems — Transactional databases, CRM, ERP, IoT devices, logs
  2. Ingestion — Extract data from sources (ETL or ELT)
  3. Storage — Land data in a data warehouse or data lake
  4. Modelling — Organise data into dimensions and measures (star schema)
  5. Visualisation — Build dashboards and reports (Power BI)

Priya’s pipeline at FreshMart:

  1. Sales data flows from 50 store POS systems → 2. Azure Data Factory extracts and transforms it nightly → 3. Lands in a Fabric data warehouse → 4. Priya builds a star schema (stores, products, dates, sales amounts) → 5. Power BI dashboards show trends to regional managers
ℹ️ ETL vs ELT — what's the difference?

Both are ways to move data from source to analytical storage:

  • ETL (Extract, Transform, Load): Extract data, transform it (clean, reshape, aggregate) in a separate engine, then load the result into the warehouse. Traditional approach.
  • ELT (Extract, Load, Transform): Extract data, load it raw into the data lake or warehouse, then transform it inside the storage system using its compute power. Modern approach — especially with Fabric and Databricks.

Key difference: In ELT, the raw data is preserved in the lake, so you can re-transform it later if business requirements change. In ETL, only the transformed result is stored.

Key concepts in analytical workloads

Data warehouses

A data warehouse is a centralised repository of structured, historical data optimised for analytical queries. Data is cleaned, transformed, and organised before loading.

Data lakes

A data lake stores raw data in its original format — structured, semi-structured, and unstructured. It’s a landing zone for everything. Analytics tools query the lake directly.

Lakehouses

A lakehouse combines the best of both — the raw storage of a data lake with the query performance and structure of a data warehouse. Microsoft Fabric uses this model.

💡 Exam tip: OLAP scenario recognition

Look for these signals in exam questions:

  • “Management wants a report on last quarter’s performance” → Analytical
  • “The dashboard shows sales trends by region” → Analytical
  • “Data is aggregated from multiple source systems” → Analytical
  • “A star schema organises data for reporting” → Analytical

If the scenario mentions “recording a sale,” “placing an order,” or “updating a record” — that’s transactional.

Flashcards

Question

What does OLAP stand for?

Click or press Enter to reveal answer

Answer

Online Analytical Processing. OLAP systems process large volumes of historical data to find patterns, trends, and insights — answering questions like 'what happened' and 'why.'

Click to flip back

Question

What's the difference between ETL and ELT?

Click or press Enter to reveal answer

Answer

ETL transforms data before loading it into storage. ELT loads raw data first, then transforms it inside the storage system. ELT preserves raw data and leverages modern compute power.

Click to flip back

Question

What is a lakehouse?

Click or press Enter to reveal answer

Answer

A lakehouse combines the raw storage flexibility of a data lake with the structured query performance of a data warehouse. Microsoft Fabric uses this approach with OneLake.

Click to flip back

Question

What is denormalised data, and why is it used in analytics?

Click or press Enter to reveal answer

Answer

Denormalised data combines related information into fewer tables (with some duplication) to reduce the number of joins needed. This speeds up read-heavy analytical queries at the cost of more storage.

Click to flip back

Knowledge check

Knowledge Check

Priya needs to create a monthly dashboard showing total sales by store, product category, and region for the past 12 months. What type of workload is this?

Knowledge Check

A company loads raw sales data into a data lake first, then uses Fabric to clean, transform, and model it for reporting. What is this approach called?

🎬 Video coming soon

Next up: Data Roles: DBA, Engineer & Analyst — meet the people who build and manage these data systems.

← Previous

Transactional Workloads: Keeping Data Consistent

Next →

Data Roles: DBA, Engineer & Analyst

Guided

I learn, I simplify, I share.

A Guide to Cloud YouTube Feedback

© 2026 Sutheesh. All rights reserved.

Guided is an independent study resource and is not affiliated with, endorsed by, or officially connected to Microsoft. Microsoft, Azure, and related trademarks are property of Microsoft Corporation. Always verify information against Microsoft Learn.