πŸ”’ Guided

Pre-launch preview. Authorised access only.

Incorrect code

Guided by A Guide to Cloud
Explore AB-900 AI-901
Guided AB-100 Domain 2
Domain 2 β€” Module 7 of 9 78%
14 of 29 overall

AB-100 Study Guide

Domain 1: Plan AI-Powered Business Solutions

  • Agent Requirements & Data Readiness
  • AI Strategy & the Cloud Adoption Framework
  • Multi-Agent Solution Design
  • Build, Buy, or Extend
  • Generative AI, Knowledge Sources & Prompt Engineering
  • Small Language Models & Model Selection
  • ROI, TCO & Business Case Analysis

Domain 2: Design AI-Powered Business Solutions

  • Copilot in D365 Customer Experience & Service
  • Agent Types: Task, Autonomous & Prompt/Response
  • Foundry Tools & Code-First Solutions
  • Copilot Studio: Topics, Flows & Prompt Actions
  • Power Apps, WAF & Data Processing
  • Extensibility: Custom Models, M365 Agents & Copilot Studio
  • MCP, Computer Use & Agent Behaviours
  • M365 Agents: Teams, SharePoint & Sales/Service in M365 Copilot
  • D365 AI Orchestration: Finance, SCM & Customer Experience

Domain 3: Deploy AI-Powered Business Solutions

  • Agent Monitoring: Tools, Metrics, and Processes
  • Telemetry Interpretation and Agent Tuning
  • Testing Strategy for AI Agents
  • Custom Model Validation and Prompt Best Practices
  • End-to-End Testing for Multi-App AI Solutions
  • ALM Foundations & Data Lifecycle for AI
  • ALM for Copilot Studio Agents
  • ALM for Microsoft Foundry Agents
  • ALM for D365 AI Features
  • Agent Security Free
  • Governance for AI Agents Free
  • Prompt Security & AI Vulnerabilities Free
  • Responsible AI & Audit Trails Free

AB-100 Study Guide

Domain 1: Plan AI-Powered Business Solutions

  • Agent Requirements & Data Readiness
  • AI Strategy & the Cloud Adoption Framework
  • Multi-Agent Solution Design
  • Build, Buy, or Extend
  • Generative AI, Knowledge Sources & Prompt Engineering
  • Small Language Models & Model Selection
  • ROI, TCO & Business Case Analysis

Domain 2: Design AI-Powered Business Solutions

  • Copilot in D365 Customer Experience & Service
  • Agent Types: Task, Autonomous & Prompt/Response
  • Foundry Tools & Code-First Solutions
  • Copilot Studio: Topics, Flows & Prompt Actions
  • Power Apps, WAF & Data Processing
  • Extensibility: Custom Models, M365 Agents & Copilot Studio
  • MCP, Computer Use & Agent Behaviours
  • M365 Agents: Teams, SharePoint & Sales/Service in M365 Copilot
  • D365 AI Orchestration: Finance, SCM & Customer Experience

Domain 3: Deploy AI-Powered Business Solutions

  • Agent Monitoring: Tools, Metrics, and Processes
  • Telemetry Interpretation and Agent Tuning
  • Testing Strategy for AI Agents
  • Custom Model Validation and Prompt Best Practices
  • End-to-End Testing for Multi-App AI Solutions
  • ALM Foundations & Data Lifecycle for AI
  • ALM for Copilot Studio Agents
  • ALM for Microsoft Foundry Agents
  • ALM for D365 AI Features
  • Agent Security Free
  • Governance for AI Agents Free
  • Prompt Security & AI Vulnerabilities Free
  • Responsible AI & Audit Trails Free
Domain 2: Design AI-Powered Business Solutions Premium ⏱ ~16 min read

MCP, Computer Use & Agent Behaviours

Design agent extensibility with Model Context Protocol, automate tasks in apps and websites with Computer Use, and configure agent behaviours including reasoning mode and voice mode in Copilot Studio.

New frontiers: MCP, Computer Use, and smart behaviours

β˜• Simple explanation

Model Context Protocol (MCP) is like a universal power adapter for AI agents. Just as a travel adapter lets your laptop plug into any country’s power outlet, MCP lets your agent plug into any tool or data source that speaks the MCP protocol β€” without building a custom connector for each one.

Computer Use is like giving your agent hands and eyes. It can see what’s on a screen and interact with it β€” clicking buttons, typing text, navigating menus. Think of it as a very patient colleague who follows your instructions exactly, even in apps that have no API.

Deep reasoning is like asking your colleague to take extra time on a tough problem β€” carefully working through the steps before giving you an answer. Voice mode is like calling that colleague on the phone instead of texting them.

MCP (Model Context Protocol) is an open standard that provides a universal interface for agents to connect to external tools, resources, and prompts. In Copilot Studio, MCP connections let agents access any MCP server β€” bringing in tools, data sources, and prompt templates without building custom connectors for each integration.

Computer Use enables Copilot Studio agents to interact with applications and websites visually β€” observing the screen and performing actions (clicking, typing, scrolling). This bridges the gap for legacy systems with no API.

Agent behaviours configure how the agent reasons and communicates: deep reasoning (preview) enables enhanced multi-step analysis for complex problems, and voice mode enables spoken interaction for phone and IVR scenarios.

Model Context Protocol (MCP) in Copilot Studio

MCP is an open protocol that standardises how AI agents connect to external tools and data. Instead of building a custom connector for every integration, you connect to an MCP server that exposes tools, resources, and prompts in a standard format.

How MCP works in Copilot Studio:

  1. MCP server β€” an external service that exposes tools (functions the agent can call), resources (data the agent can read), and prompts (pre-built prompt templates)
  2. Connection β€” Copilot Studio connects to the MCP server endpoint
  3. Discovery β€” the agent discovers available tools and resources from the MCP server
  4. Invocation β€” during a conversation, the agent calls MCP tools when relevant to the user’s request
MCP vs custom connectors vs Copilot connectors
FeatureArchitectureBest ForEffort
MCP connectionsOpen protocol β€” connect to any MCP-compatible server. Agent discovers tools dynamicallyWhen MCP servers already exist for the tools you need, when connecting to open-source tool ecosystems, or when you want a single integration pattern across many toolsLow if MCP server exists. Medium if you need to build one
Custom connectorsPoint-to-point REST API integration β€” you define the schema, authentication, and endpointsWhen the external system has a REST API but no MCP server, when you need fine-grained control over the API interactionMedium β€” define connector schema, configure auth, map data
Copilot connectorsPrebuilt Microsoft connectors in the connector gallery β€” ready to use with authentication configuredWhen a prebuilt connector already exists for the target system (Salesforce, ServiceNow, SAP, etc.)Low β€” select from gallery, configure credentials
πŸ’‘ Scenario: Natalie implements MCP for a client's inventory system

Natalie’s client has a custom inventory management system built in-house. Ravi (senior developer) has already built an MCP server that exposes three tools:

  • check_stock(product_id) β€” returns current inventory levels
  • find_alternatives(product_id) β€” finds substitute products when stock is low
  • create_reorder(product_id, quantity) β€” creates a purchase order

Copilot Studio integration:

  1. Natalie adds the MCP connection in Copilot Studio, pointing to the MCP server endpoint
  2. The agent automatically discovers the three tools
  3. When a sales rep asks β€œDo we have 500 units of part X in stock?”, the agent calls check_stock and responds with current levels
  4. If stock is low, the agent proactively calls find_alternatives and offers substitutes
  5. The rep can say β€œOrder 200 more of part X” and the agent calls create_reorder

Why MCP over a custom connector? The MCP server already exists, the tool discovery is automatic, and if Ravi adds new tools later (like track_shipment), the agent discovers them without reconfiguration.

πŸ’‘ Important: MCP requires generative orchestration

MCP connections in Copilot Studio require generative orchestration to be enabled on the agent. Classic topic-based agents cannot call MCP servers directly β€” MCP tools are invoked through the generative orchestrator, which decides when and how to use them based on the conversation context.

When to prefer MCP: Choose MCP when you want centrally managed, reusable tools that multiple agents can share. An MCP server built once can serve many agents β€” unlike custom connectors that are configured per agent.

Computer Use in Copilot Studio (preview)

Computer Use gives agents the ability to interact with applications and websites by seeing the screen and performing actions β€” clicking buttons, typing into fields, navigating menus, and reading displayed content.

Preview notice: Computer Use is currently in preview and available in the US region only. It requires generative orchestration to be enabled. There are two hosting options: hosted machines (Microsoft-managed VMs β€” simpler setup, Microsoft handles infrastructure) and bring-your-own machines (self-hosted VMs β€” more control, you manage the infrastructure). Consider latency implications for production use, as each screen observation and action cycle adds delay.

Use cases:

ScenarioWhy Computer UseAlternative
Automating data entry in a legacy ERP with no APINo programmatic access β€” the only interface is the UIBuild an API layer (expensive, time-consuming)
Testing web applicationsNavigate pages, fill forms, verify contentDedicated testing tools (Playwright, Selenium)
Extracting data from desktop applicationsInformation is on-screen but not accessible via APIsManual extraction or custom screen scraping
Training and onboarding demonstrationsShow step-by-step how to use an applicationRecorded videos or documentation

Design considerations for Computer Use:

  • Latency: Screen observation and action take time β€” expect 2-5 seconds per step. Not suitable for real-time interactions
  • Accuracy: Complex UIs with dynamic elements can confuse the agent. Simple, consistent layouts work best
  • Security: The agent sees everything on screen β€” ensure sensitive data is masked or the session is scoped appropriately
  • Reliability: UI changes (button moves, new pop-up) can break automation. Build error handling and retry logic
  • Scope: Define exactly which applications and actions the agent is allowed to interact with
πŸ’‘ Scenario: Ravi builds a Computer Use agent for legacy ERP data entry

Ravi Krishnan (Cloudbridge Partners) automates data entry into a client’s legacy ERP system that has no API. The client’s warehouse team currently copies shipping data from emails into the ERP manually β€” 80 entries per day.

Computer Use agent design:

  1. Agent receives shipping notification data (structured from an email processing step)
  2. Opens the legacy ERP application
  3. Navigates to the β€œNew Shipment” form
  4. Fills in fields: tracking number, carrier, ship date, destination, weight
  5. Clicks β€œSave”
  6. Verifies the confirmation message appears
  7. Logs the result for audit

Guardrails:

  • Agent can only interact with the ERP’s shipment entry form β€” no access to other modules
  • Every entry is logged with before/after screenshots for audit
  • If the agent encounters an unexpected screen (error dialog, login prompt), it stops and alerts the warehouse manager
  • A human reviews a random 10% sample daily

Natalie presents the ROI: 80 entries at 3 minutes each = 4 hours of manual work per day. The agent handles 90% automatically, saving 3.6 hours daily.

Deep reasoning (preview)

Deep reasoning enables enhanced multi-step analysis for complex tasks. When enabled, the agent methodically breaks problems into steps and plans its approach before responding. This produces higher-quality answers for complex queries but is slower and costlier than standard orchestration.

When to enable deep reasoning:

  • Multi-step calculations or analyses
  • Complex decision-making with multiple variables
  • Tasks that require comparing options and making trade-offs
  • Situations requiring careful, methodical problem decomposition

When NOT to enable deep reasoning:

  • Simple Q&A from a knowledge base
  • Straightforward topic routing
  • Tasks where speed matters more than depth

Note: Deep reasoning improves the agent’s internal analysis quality. It does not expose its reasoning chain to the end user β€” the user receives a final answer, not the intermediate steps.

Voice mode

Voice mode enables spoken interaction β€” the agent listens to speech, processes it, and responds with synthesised voice. Designed for phone and IVR (Interactive Voice Response) scenarios.

Design considerations for voice agents:

  • Latency is critical β€” voice users expect fast responses. Keep prompts concise and responses under 30 seconds
  • No visual fallback β€” you can’t show tables, links, or images. Design responses for audio only
  • Confirmation patterns β€” β€œI heard you say you want to cancel order 1234. Is that correct?”
  • Interruption handling β€” users may interrupt mid-response. Design for barge-in support
  • Escalation β€” always provide a β€œspeak to a human” option

Flashcards

Question

What is MCP (Model Context Protocol) and how does it work in Copilot Studio?

Click or press Enter to reveal answer

Answer

MCP is an open protocol that provides a standard interface for agents to connect to external tools, resources, and prompts. In Copilot Studio, you add an MCP connection to a server endpoint β€” the agent automatically discovers available tools and can invoke them during conversations. It replaces the need for individual custom connectors when MCP servers exist.

Click to flip back

Question

What is Computer Use in Copilot Studio?

Click or press Enter to reveal answer

Answer

Computer Use (preview, US region only) enables agents to interact with applications and websites visually β€” observing the screen and performing actions like clicking, typing, and navigating. It bridges the gap for legacy systems with no API. Requires generative orchestration. Hosting options: hosted machines (Microsoft-managed) or bring-your-own machines (self-hosted). Design considerations: latency (2-5 seconds per step), accuracy on complex UIs, security (agent sees the screen), and reliability (UI changes can break automation).

Click to flip back

Question

When should you enable deep reasoning for a Copilot Studio agent?

Click or press Enter to reveal answer

Answer

Enable deep reasoning (preview) for complex tasks: multi-step calculations, decision-making with multiple variables, and comparing options with trade-offs. Deep reasoning improves internal analysis quality but is slower and costlier than standard orchestration β€” the user receives a final answer, not intermediate steps. Don't enable it for simple Q&A or straightforward routing.

Click to flip back

Question

What are the key design considerations for voice-enabled agents?

Click or press Enter to reveal answer

Answer

Latency (voice users expect fast responses), no visual fallback (design for audio only β€” no tables or links), confirmation patterns (repeat back what the user said), interruption handling (barge-in support), and always provide escalation to a human agent.

Click to flip back

Question

When should you choose MCP over a custom connector in Copilot Studio?

Click or press Enter to reveal answer

Answer

Choose MCP when: an MCP server already exists for the target tool, you want to connect to an open-source tool ecosystem, or you want dynamic tool discovery (agent finds new tools automatically). Choose custom connectors when: no MCP server exists, you need fine-grained control over API interactions, or the external system only has a REST API.

Click to flip back

Knowledge check

Knowledge Check

A client has a 15-year-old desktop HR application with no API. HR staff spend 2 hours daily copying new hire data from a spreadsheet into the application's forms. Natalie proposes using Computer Use to automate this. What is the MOST critical design element she must address?

Knowledge Check

Ravi has built an MCP server that exposes 5 tools for a client's custom CRM. He adds the MCP connection in Copilot Studio. Two weeks later, he adds 2 new tools to the MCP server. What does the Copilot Studio agent need?

Knowledge Check

Jordan is designing a patient appointment scheduling agent for CareFirst Health. Patients will call a phone number, speak their scheduling request, and the agent should book, reschedule, or cancel appointments. Which combination of agent behaviours should Jordan enable?

🎬 Video coming soon

Next up: M365 Agents: Teams, SharePoint & Sales/Service in M365 Copilot β€” optimising agent design across Microsoft 365, configuring Sales and Service in M365 Copilot, and leveraging the AI hub in Power Platform.

← Previous

Extensibility: Custom Models, M365 Agents & Copilot Studio

Next β†’

M365 Agents: Teams, SharePoint & Sales/Service in M365 Copilot

Guided

I learn, I simplify, I share.

A Guide to Cloud YouTube Feedback

© 2026 Sutheesh. All rights reserved.

Guided is an independent study resource and is not affiliated with, endorsed by, or officially connected to Microsoft. Microsoft, Azure, and related trademarks are property of Microsoft Corporation. Always verify information against Microsoft Learn.