πŸ”’ Guided

Pre-launch preview. Authorised access only.

Incorrect code

Guided by A Guide to Cloud
Explore AB-900 AI-901
Guided AZ-305 Domain 2
Domain 2 β€” Module 6 of 6 100%
13 of 30 overall

AZ-305 Study Guide

Domain 1: Design Identity, Governance, and Monitoring Solutions

  • Monitoring & Logging Design
  • Choosing Authentication Methods
  • Designing Identity Management
  • Authorizing Access to Resources
  • Secrets, Keys & Certificates
  • Governance at Scale
  • Compliance & Identity Governance

Domain 2: Design Data Storage Solutions

  • Relational Data: Choosing Your SQL Platform
  • Database Performance & Scalability
  • Database Security & Compliance
  • Cosmos DB & Semi-Structured Data
  • Blob, Data Lake & Azure Files
  • Data Integration & Analytics

Domain 3: Design Business Continuity Solutions

  • Recovery Objectives: RPO, RTO & SLA Free
  • Backup & Recovery for Compute Free
  • Backup for Databases & Unstructured Data Free
  • High Availability for Compute Free
  • High Availability for Data Free

Domain 4: Design Infrastructure Solutions

  • Compute Design: VMs & When to Use Them
  • Container Solutions: AKS, ACI & Container Apps
  • Serverless & Batch Processing
  • Messaging Architecture
  • Event-Driven Architecture
  • API Integration & Caching
  • App Configuration & Automated Deployment
  • Migration Strategy & Assessment
  • Executing Migrations
  • Network Connectivity: Internet & Hybrid
  • Network Security & Performance
  • Load Balancing & Routing

AZ-305 Study Guide

Domain 1: Design Identity, Governance, and Monitoring Solutions

  • Monitoring & Logging Design
  • Choosing Authentication Methods
  • Designing Identity Management
  • Authorizing Access to Resources
  • Secrets, Keys & Certificates
  • Governance at Scale
  • Compliance & Identity Governance

Domain 2: Design Data Storage Solutions

  • Relational Data: Choosing Your SQL Platform
  • Database Performance & Scalability
  • Database Security & Compliance
  • Cosmos DB & Semi-Structured Data
  • Blob, Data Lake & Azure Files
  • Data Integration & Analytics

Domain 3: Design Business Continuity Solutions

  • Recovery Objectives: RPO, RTO & SLA Free
  • Backup & Recovery for Compute Free
  • Backup for Databases & Unstructured Data Free
  • High Availability for Compute Free
  • High Availability for Data Free

Domain 4: Design Infrastructure Solutions

  • Compute Design: VMs & When to Use Them
  • Container Solutions: AKS, ACI & Container Apps
  • Serverless & Batch Processing
  • Messaging Architecture
  • Event-Driven Architecture
  • API Integration & Caching
  • App Configuration & Automated Deployment
  • Migration Strategy & Assessment
  • Executing Migrations
  • Network Connectivity: Internet & Hybrid
  • Network Security & Performance
  • Load Balancing & Routing
Domain 2: Design Data Storage Solutions Premium ⏱ ~18 min read

Data Integration & Analytics

Azure Data Factory, Synapse Analytics, Microsoft Fabric, and Stream Analytics β€” design data pipelines that move, transform, and analyse data across your Azure estate.

Why data integration design matters

β˜• Simple explanation

Data is useless if it’s stuck in silos. Integration means connecting data from different sources (databases, APIs, files, streams), transforming it into a useful shape, and delivering it to where analysis happens.

Two patterns: Batch integration (move yesterday’s data overnight β€” Data Factory) and Real-time streaming (process events as they happen β€” Stream Analytics, Event Hubs).

For analysis, think Synapse Analytics for big data warehousing, Microsoft Fabric as the all-in-one analytics platform, and Power BI for business reporting.

Data integration and analytics architecture covers:

  • Batch ETL/ELT: Azure Data Factory pipelines β€” orchestrate data movement and transformation
  • Real-time streaming: Stream Analytics, Event Hubs β€” process events with sub-second latency
  • Data warehousing: Synapse Analytics dedicated/serverless SQL pools
  • Unified analytics: Microsoft Fabric β€” combines Data Factory, Synapse, Power BI into one SaaS platform
  • Reporting: Power BI for business intelligence dashboards and self-service analytics

Data integration services

Data Integration Service Comparison
ServicePatternData VolumeLatencyBest For
Azure Data FactoryBatch ETL/ELTLarge (TB+)Minutes to hoursScheduled data movement and transformation pipelines
Synapse PipelinesBatch ETL/ELT (same engine as ADF)Large (TB+)Minutes to hoursWhen analytics and integration are in the same workspace
Azure Stream AnalyticsReal-time stream processingContinuousSub-second to secondsIoT telemetry, real-time dashboards, event-driven alerts
Azure Logic AppsWorkflow automationSmall to mediumSeconds to minutesAPI integration, business workflows, event-triggered actions
Azure Event HubsEvent ingestion (not processing)Millions of events/secMilliseconds (ingestion)High-throughput event ingestion before processing

ETL vs ELT

PatternHow It WorksBest For
ETL (Extract, Transform, Load)Transform data BEFORE loading into destinationWhen destination has limited compute (smaller databases)
ELT (Extract, Load, Transform)Load raw data first, transform IN the destinationWhen destination has powerful compute (Synapse, Databricks, Fabric)

πŸ—οΈ Priya’s integration design: GlobalTech uses ELT pattern:

  1. Extract: Data Factory copies raw data from 15 source systems (SQL, SAP, files)
  2. Load: Raw data lands in ADLS Gen2 (bronze/raw layer)
  3. Transform: Synapse Spark transforms raw β†’ curated (silver) β†’ aggregated (gold)
  4. Serve: Power BI connects to the gold layer for executive dashboards

Data analytics services

Analytics Platform Comparison
ServiceTypeCompute ModelBest For
Synapse Dedicated SQL PoolData warehouseProvisioned (DWU β€” always-on)Large-scale, predictable warehouse workloads
Synapse Serverless SQL PoolQuery-on-demandServerless (pay per query)Ad-hoc queries on data lake files without loading
Synapse SparkBig data processingSpark clusters (auto-scale)Data engineering, ML, complex transformations
Microsoft FabricUnified analytics SaaSCapacity-basedAll-in-one: ingestion, transformation, warehouse, reporting
Azure DatabricksSpark-based analyticsClusters (auto-scale)Advanced ML, data science, delta lake architecture
Power BIBusiness intelligenceCapacity or Pro licensesDashboards, reports, self-service analytics

πŸš€ Marcus’s analytics choice: NovaSaaS adopted Microsoft Fabric for their analytics stack:

  • OneLake as the unified data lake (replacing separate ADLS accounts)
  • Data Factory in Fabric for ingestion pipelines
  • Fabric Data Warehouse for SQL-based analytics (serverless, no cluster management)
  • Power BI embedded in their SaaS product for customer-facing dashboards
ℹ️ Design decision: Synapse vs Fabric vs Databricks

Choose Synapse Analytics when:

  • You need dedicated SQL pool for predictable data warehouse workloads
  • You’re already invested in the Synapse ecosystem
  • You need Spark AND SQL in one workspace

Choose Microsoft Fabric when:

  • You want a unified SaaS platform (no infrastructure management)
  • Your team uses Power BI heavily (Fabric integrates natively)
  • You want OneLake as a single data lake across all workloads

Choose Azure Databricks when:

  • Advanced ML/data science is the primary workload
  • You use Delta Lake architecture
  • You need Spark ecosystem tools (MLflow, Delta Live Tables)
  • Multi-cloud portability matters (Databricks runs on AWS/GCP too)

Real-time streaming architecture

For scenarios needing sub-second processing:

Event Sources β†’ Event Hubs (ingestion) β†’ Stream Analytics (processing) β†’ Outputs
ComponentRoleScale
Event HubsIngestion buffer β€” receives millions of events/secPartition-based, auto-inflate
Stream AnalyticsSQL-like queries on streams β€” windowing, aggregation, joinsStreaming Units (auto-scale)
OutputsCosmos DB, SQL, Blob, Power BI, FunctionsMultiple simultaneous outputs

🏦 Elena’s real-time fraud detection:

  • Card transactions β†’ Event Hubs (millions/second ingestion)
  • Stream Analytics applies fraud rules (unusual amounts, foreign locations, velocity checks)
  • Suspicious transactions β†’ Cosmos DB for investigation dashboard
  • Alerts β†’ Azure Functions β†’ notify fraud team via Teams

Knowledge check

Question

What's the difference between ETL and ELT?

Click or press Enter to reveal answer

Answer

ETL transforms data before loading (transform happens in transit). ELT loads raw data first, then transforms in the destination. ELT is preferred when the destination has powerful compute (Synapse, Fabric, Databricks) β€” it's more scalable and keeps raw data available for re-processing.

Click to flip back

Question

When should you recommend Microsoft Fabric over Azure Synapse?

Click or press Enter to reveal answer

Answer

Fabric is the newer, unified SaaS platform β€” choose it when you want: no infrastructure management, OneLake as a unified data lake, native Power BI integration, and an all-in-one experience. Synapse is better when you need dedicated SQL pools for heavy warehouse workloads or are already invested in the Synapse ecosystem.

Click to flip back

Question

What Azure service processes millions of events per second for real-time analytics?

Click or press Enter to reveal answer

Answer

Azure Event Hubs for ingestion (receives events) + Azure Stream Analytics for processing (SQL-like queries on streams). Event Hubs is the buffer, Stream Analytics is the processor. Outputs go to Cosmos DB, SQL, Power BI, or other services.

Click to flip back

Knowledge Check

πŸ—οΈ GlobalTech needs to consolidate data from 15 source systems (SQL Server, SAP, CSV files) into a data lake for analytics. Data volumes are 2 TB daily. The analytics team wants to run Spark transformations and Power BI reports. Which architecture should Priya recommend?

Knowledge Check

🏦 FinSecure Bank processes 10 million card transactions per hour. They need real-time fraud detection with sub-second alerting. Suspicious transactions must be stored for investigation. Which architecture should Elena recommend?

🎬 Video coming soon


Domain 2 complete! You’ve designed relational databases, NoSQL with Cosmos DB, unstructured storage, and data integration pipelines.

Next up: Now let’s design for when things go wrong β€” Recovery Objectives: RPO, RTO & SLA.

← Previous

Blob, Data Lake & Azure Files

Next β†’

Recovery Objectives: RPO, RTO & SLA

Guided

I learn, I simplify, I share.

A Guide to Cloud YouTube Feedback

© 2026 Sutheesh. All rights reserved.

Guided is an independent study resource and is not affiliated with, endorsed by, or officially connected to Microsoft. Microsoft, Azure, and related trademarks are property of Microsoft Corporation. Always verify information against Microsoft Learn.