πŸ”’ Guided

Pre-launch preview. Authorised access only.

Incorrect code

Guided by A Guide to Cloud
Explore AB-900 AI-901
Guided AZ-140 Domain 1
Domain 1 β€” Module 8 of 12 67%
8 of 28 overall

AZ-140 Study Guide

Domain 1: Plan and Implement an AVD Infrastructure

  • AVD Architecture: The Big Picture Free
  • Network Capacity and Design Free
  • RDP Shortpath, Multipath and QoS Free
  • Private Link and Network Troubleshooting Free
  • Storage Planning for User Data Free
  • File Shares and Azure NetApp Files Free
  • Host Pool Architecture: Personal vs Pooled Free
  • Sizing for Performance and Capacity Free
  • Creating Host Pools and Session Hosts Free
  • Session Host Licensing Free
  • Building Session Host Images Free
  • Image Lifecycle and Compute Gallery Free

Domain 2: Plan and Implement Identity and Security

  • Identity Scenarios for AVD
  • RBAC, Conditional Access and SSO
  • Defending AVD with Microsoft Defender
  • Network Security: NSGs, Firewall, Bastion
  • Threat Protection and Confidential VMs

Domain 3: Plan and Implement User Environments and Apps

  • FSLogix Profile Containers and ODFC
  • FSLogix Cloud Cache and Application Masking
  • AVD Clients: Choose and Deploy
  • User Experience and Session Settings
  • Application Groups and RemoteApp
  • Microsoft 365, Teams and OneDrive on AVD
  • App Attach: Dynamic Application Delivery

Domain 4: Monitor and Maintain an AVD Infrastructure

  • Monitoring AVD with Azure Monitor
  • Autoscaling and Session Management
  • Update Strategy and Backups
  • Disaster Recovery and Multi-Region

AZ-140 Study Guide

Domain 1: Plan and Implement an AVD Infrastructure

  • AVD Architecture: The Big Picture Free
  • Network Capacity and Design Free
  • RDP Shortpath, Multipath and QoS Free
  • Private Link and Network Troubleshooting Free
  • Storage Planning for User Data Free
  • File Shares and Azure NetApp Files Free
  • Host Pool Architecture: Personal vs Pooled Free
  • Sizing for Performance and Capacity Free
  • Creating Host Pools and Session Hosts Free
  • Session Host Licensing Free
  • Building Session Host Images Free
  • Image Lifecycle and Compute Gallery Free

Domain 2: Plan and Implement Identity and Security

  • Identity Scenarios for AVD
  • RBAC, Conditional Access and SSO
  • Defending AVD with Microsoft Defender
  • Network Security: NSGs, Firewall, Bastion
  • Threat Protection and Confidential VMs

Domain 3: Plan and Implement User Environments and Apps

  • FSLogix Profile Containers and ODFC
  • FSLogix Cloud Cache and Application Masking
  • AVD Clients: Choose and Deploy
  • User Experience and Session Settings
  • Application Groups and RemoteApp
  • Microsoft 365, Teams and OneDrive on AVD
  • App Attach: Dynamic Application Delivery

Domain 4: Monitor and Maintain an AVD Infrastructure

  • Monitoring AVD with Azure Monitor
  • Autoscaling and Session Management
  • Update Strategy and Backups
  • Disaster Recovery and Multi-Region
Domain 1: Plan and Implement an AVD Infrastructure Free ⏱ ~14 min read

Sizing for Performance and Capacity

Choose the right VM families, calculate vCPU-to-user ratios, plan GPU workloads, and select optimal disk types for AVD session hosts.

Right-sizing your session hosts

β˜• Simple explanation

Imagine each session host VM as a bus.

A small bus (4 vCPU) can carry a few passengers (users). A large bus (16 vCPU) carries more. If the passengers only carry handbags (light workload β€” email and web), you fit lots on one bus. If they bring full suitcases (heavy workload β€” CAD software), you need fewer passengers per bus or a bigger bus. Picking the wrong bus size means either wasted money (huge empty bus) or angry passengers (overloaded, slow bus).

VM sizing directly impacts user experience and monthly cost. Under-sized VMs cause slow application response and sign-in delays. Over-sized VMs waste budget. Microsoft publishes sizing guidelines based on workload type, and the AZ-140 exam expects you to apply them to real scenarios β€” calculating how many VMs you need for a given user count and workload.

Workload types and VM sizing

Microsoft defines four workload types. The exam uses these categories in scenario questions:

WorkloadExamplesMax users per vCPUMin VM specExample VM
LightData entry, basic web browsing68 vCPU, 16 GB RAMD8s_v5
MediumOffice apps, email, light databases48 vCPU, 16 GB RAMD8s_v5
HeavySoftware development, content creation28 vCPU, 16 GB RAMD8s_v5
PowerCAD, 3D modelling, simulation, video editing18 vCPU, 16 GB RAM, GPUNV-series

Reading the table: β€œMax users per vCPU” tells you the density. A D8s_v5 with 8 vCPUs can host up to 48 light users (8 times 6), 32 medium users (8 times 4), or 16 heavy users (8 times 2). Power users usually get 1:1 or at most 2:1 ratios.

🌐 Priya at NomadTech: β€œWe profiled our 200 users: 150 are light (email, browser, Teams), 45 are medium (Excel dashboards, Power BI), and 5 are power (video editing). That profile drove completely different VM sizes across our three pools.”

VM families for AVD

VM families for AVD workloads
VM FamilyOptimised forTypical AVD useExample sizes
D-series v5General purpose (balanced CPU/memory)Default choice for light and medium workloadsD4s_v5 (4 vCPU/16 GB), D8s_v5 (8/32), D16s_v5 (16/64)
E-series v5Memory optimisedHeavy workloads with large datasets, big Excel files, in-memory databasesE4s_v5 (4 vCPU/32 GB), E8s_v5 (8/64)
F-series v2Compute optimisedCPU-intensive light workloads (high density)F8s_v2 (8 vCPU/16 GB)
NV-seriesGPU β€” virtualised (vGPU)Graphics rendering, CAD, video playbackNV6ads_A10_v5 (6 vCPU/55 GB/A10 GPU partition)
NC-seriesGPU β€” full GPU (CUDA/ML)Machine learning inference, scientific simulationsNC4as_T4_v3 (4 vCPU/28 GB/T4 GPU)
B-seriesBurstable (credits)Dev/test or very light intermittent workloads onlyB4ms (4 vCPU/16 GB)

Key rules:

  • D-series is the default β€” start here unless you have a specific reason not to
  • E-series when users complain about RAM pressure (lots of browser tabs, big spreadsheets)
  • NV-series for graphics-accelerated desktops (CAD, video, 3D)
  • B-series only for dev/test β€” burstable VMs throttle once credits run out, making them unsuitable for production workloads

🏒 Raj’s CAD team: β€œWe tested D16s_v5 for CAD but Revit rendering was painfully slow without GPU acceleration. We switched to NV-series with a partitioned NVIDIA A10 GPU. Now each designer gets a slice of GPU and render times dropped by 80%. Cost went up, but productivity went up even more.”

Memory planning

While vCPU ratios get the most attention, memory is often the real bottleneck:

WorkloadRAM per user (guideline)
Light2 GB
Medium3-4 GB
Heavy6-8 GB
Power8-16+ GB

A D8s_v5 has 32 GB RAM. With 32 medium users (per the 4:1 vCPU ratio), that is only 1 GB per user β€” well below the 3-4 GB guideline. You may need to drop density or choose a memory-optimised E-series VM.

πŸ’‘ Exam tip: RAM vs vCPU β€” which limits first?

The exam may give you a VM size and a user count and ask if it is adequate. Always check BOTH vCPU ratio and memory per user. A D8s_v5 (8 vCPU, 32 GB) hosting 32 medium users satisfies the 4:1 vCPU ratio but only gives 1 GB RAM per user (well below the 3-4 GB recommendation). The correct answer would be to either reduce density or switch to E8s_v5 (8 vCPU, 64 GB) for 2 GB per user.

GPU workloads

Some users need hardware graphics acceleration. AVD supports GPU-enabled VMs for:

  • RemoteFX vGPU β€” shares a physical GPU among multiple users (NV-series)
  • Full GPU pass-through β€” dedicates the entire GPU to one user (NC-series for compute, NV for graphics)

Use cases: CAD/CAM (AutoCAD, Revit, SolidWorks), video editing (Premiere, DaVinci), 3D visualisation (Power BI 3D visuals), medical imaging

🎧 Mia at Horizons Health: β€œDr. Patel’s radiology workstation needs GPU for DICOM 3D rendering. We deployed NV-series personal desktops for the 8 radiologists. Everyone else in the clinics uses D-series pooled desktops for EHR and scheduling β€” no GPU needed.”

OS disk types

The operating system disk impacts boot time and application launch speed:

Disk typePerformanceCostPersistent?Best for
Standard HDDLow IOPS, high latencyLowestYesDev/test only
Standard SSDModerate IOPSLowYesLight workloads on a budget
Premium SSDHigh IOPS, low latencyMediumYesProduction personal desktops
Ephemeral OS diskVery high (uses local VM cache)Included in VM priceNo β€” resets on reimage/deallocateProduction pooled desktops

Ephemeral OS disks

Ephemeral disks use the VM’s local temporary storage (or cache) instead of a remote managed disk. Benefits:

  • Faster boot and reimage β€” no network storage latency
  • No disk cost β€” the local cache is included in the VM price
  • Stateless by design β€” perfect for pooled pools where VMs are reimaged regularly

The tradeoff: data on the OS disk is lost when the VM is deallocated or reimaged. This is ideal for pooled host pools (user data is in FSLogix on the file share, not on the OS disk) but NOT suitable for personal desktops where users install software locally.

πŸ’‘ Exam tip: Ephemeral disk placement

Ephemeral OS disks can be placed on the VM cache or the temporary (temp) disk. Cache placement gives better IOPS but only works if the VM size has enough cache capacity for the OS image. The exam may ask where to place the ephemeral disk β€” check that the VM size supports it. For example, D8s_v5 has a cache size of 200 GiB, which comfortably fits a standard Windows image.

Session limits

Each session host has a maximum session limit that caps how many users can connect to it:

  • For breadth-first load balancing, the default max session limit is 999,999 (effectively unlimited β€” relies on even distribution)
  • For depth-first, you MUST set a realistic max session limit (e.g., 12 users per D8s_v5 for medium workloads)
  • The max session limit should be based on your vCPU and memory calculations

If all session hosts reach their max session limit, new users cannot connect. You need to either increase the limit or add more session hosts.

Capacity planning walkthrough

🏒 Raj’s calculation for TerraStack:

Scenario: 500 medium-workload users, breadth-first load balancing, target 4:1 vCPU ratio, 4 GB RAM per user.

  1. Pick VM size: D16s_v5 β€” 16 vCPU, 64 GB RAM
  2. vCPU capacity: 16 vCPU times 4 users per vCPU = 64 users per VM
  3. Memory check: 64 GB divided by 64 users = 1 GB per user β€” too low for medium workloads
  4. Adjust for memory: 64 GB divided by 4 GB per user = 16 users per VM
  5. VMs needed: 500 users divided by 16 per VM = 31.25, round up to 32 VMs
  6. Add buffer: 20% buffer for failover and updates = 32 times 1.2 = 38.4, round up to 39 VMs

Notice how memory (step 4) limited capacity far more than vCPU ratio alone (step 2). Always check both.

🎧 Mia’s clinic sizing: β€œWe have 400 light-workload clinical staff (EHR and scheduling). With D8s_v5 (8 vCPU, 32 GB), we get 48 users per VM from the vCPU ratio (8 times 6). Memory check: 32 GB divided by 48 = 0.67 GB β€” borderline for light. We settled on 24 users per VM, which gives 1.3 GB RAM each, and that needs 17 VMs plus a 20% buffer = 21 VMs total.”

Question

What is the maximum users-per-vCPU ratio for light workloads?

Click or press Enter to reveal answer

Answer

6 users per vCPU. An 8-vCPU VM can theoretically host up to 48 light users, but always verify that memory per user is adequate too.

Click to flip back

Question

Which VM family should you choose for CAD and 3D rendering workloads?

Click or press Enter to reveal answer

Answer

NV-series (GPU-enabled with virtualised GPU). For dedicated GPU compute (ML inference), use NC-series.

Click to flip back

Question

Why are ephemeral OS disks ideal for pooled host pools?

Click or press Enter to reveal answer

Answer

They use local VM cache (no disk cost, very fast boot and reimage), and since pooled VMs are stateless β€” user data lives in FSLogix on the file share β€” losing OS disk data on reimage is expected behaviour.

Click to flip back

Question

What should you check beyond vCPU ratio when sizing VMs?

Click or press Enter to reveal answer

Answer

Memory per user. The vCPU ratio may allow 64 users on a D16s_v5, but 64 GB divided by 64 users is only 1 GB per user β€” well below the 3-4 GB needed for medium workloads.

Click to flip back

Question

What percentage buffer should you add for failover and updates in capacity planning?

Click or press Enter to reveal answer

Answer

At least 20%. This accounts for VMs being unavailable during OS patching, AVD agent updates, or unexpected failures.

Click to flip back

Knowledge Check

Priya has 150 light-workload users and 45 medium-workload users. She is using D8s_v5 VMs (8 vCPU, 32 GB RAM). How many VMs does she need for the medium-workload pool (including 20% buffer)?

Knowledge Check

Raj's CAD team of 50 users runs AutoCAD and Revit with heavy 3D rendering. Which VM configuration is most appropriate?

Knowledge Check

Which OS disk type should you use for a production pooled host pool where VMs are reimaged weekly?


Next up: Time to actually deploy the infrastructure β€” Creating Host Pools and Session Hosts.

🎬 Video coming soon

VM Sizing and Capacity Planning for AVD

← Previous

Host Pool Architecture: Personal vs Pooled

Next β†’

Creating Host Pools and Session Hosts

Guided

I learn, I simplify, I share.

A Guide to Cloud YouTube Feedback

© 2026 Sutheesh. All rights reserved.

Guided is an independent study resource and is not affiliated with, endorsed by, or officially connected to Microsoft. Microsoft, Azure, and related trademarks are property of Microsoft Corporation. Always verify information against Microsoft Learn.