🔒 Guided

Pre-launch preview. Authorised access only.

Incorrect code

Guided by A Guide to Cloud
Explore AB-900 AI-901
Guided AB-100 Domain 3
Domain 3 — Module 1 of 13 8%
17 of 29 overall

AB-100 Study Guide

Domain 1: Plan AI-Powered Business Solutions

  • Agent Requirements & Data Readiness
  • AI Strategy & the Cloud Adoption Framework
  • Multi-Agent Solution Design
  • Build, Buy, or Extend
  • Generative AI, Knowledge Sources & Prompt Engineering
  • Small Language Models & Model Selection
  • ROI, TCO & Business Case Analysis

Domain 2: Design AI-Powered Business Solutions

  • Copilot in D365 Customer Experience & Service
  • Agent Types: Task, Autonomous & Prompt/Response
  • Foundry Tools & Code-First Solutions
  • Copilot Studio: Topics, Flows & Prompt Actions
  • Power Apps, WAF & Data Processing
  • Extensibility: Custom Models, M365 Agents & Copilot Studio
  • MCP, Computer Use & Agent Behaviours
  • M365 Agents: Teams, SharePoint & Sales/Service in M365 Copilot
  • D365 AI Orchestration: Finance, SCM & Customer Experience

Domain 3: Deploy AI-Powered Business Solutions

  • Agent Monitoring: Tools, Metrics, and Processes
  • Telemetry Interpretation and Agent Tuning
  • Testing Strategy for AI Agents
  • Custom Model Validation and Prompt Best Practices
  • End-to-End Testing for Multi-App AI Solutions
  • ALM Foundations & Data Lifecycle for AI
  • ALM for Copilot Studio Agents
  • ALM for Microsoft Foundry Agents
  • ALM for D365 AI Features
  • Agent Security Free
  • Governance for AI Agents Free
  • Prompt Security & AI Vulnerabilities Free
  • Responsible AI & Audit Trails Free

AB-100 Study Guide

Domain 1: Plan AI-Powered Business Solutions

  • Agent Requirements & Data Readiness
  • AI Strategy & the Cloud Adoption Framework
  • Multi-Agent Solution Design
  • Build, Buy, or Extend
  • Generative AI, Knowledge Sources & Prompt Engineering
  • Small Language Models & Model Selection
  • ROI, TCO & Business Case Analysis

Domain 2: Design AI-Powered Business Solutions

  • Copilot in D365 Customer Experience & Service
  • Agent Types: Task, Autonomous & Prompt/Response
  • Foundry Tools & Code-First Solutions
  • Copilot Studio: Topics, Flows & Prompt Actions
  • Power Apps, WAF & Data Processing
  • Extensibility: Custom Models, M365 Agents & Copilot Studio
  • MCP, Computer Use & Agent Behaviours
  • M365 Agents: Teams, SharePoint & Sales/Service in M365 Copilot
  • D365 AI Orchestration: Finance, SCM & Customer Experience

Domain 3: Deploy AI-Powered Business Solutions

  • Agent Monitoring: Tools, Metrics, and Processes
  • Telemetry Interpretation and Agent Tuning
  • Testing Strategy for AI Agents
  • Custom Model Validation and Prompt Best Practices
  • End-to-End Testing for Multi-App AI Solutions
  • ALM Foundations & Data Lifecycle for AI
  • ALM for Copilot Studio Agents
  • ALM for Microsoft Foundry Agents
  • ALM for D365 AI Features
  • Agent Security Free
  • Governance for AI Agents Free
  • Prompt Security & AI Vulnerabilities Free
  • Responsible AI & Audit Trails Free
Domain 3: Deploy AI-Powered Business Solutions Premium ⏱ ~14 min read

Agent Monitoring: Tools, Metrics, and Processes

Recommend monitoring processes and tools, and track agent performance metrics for AI-powered business solutions.

Agent Monitoring: Tools, Metrics, and Processes

☕ Simple explanation

Think of agent monitoring like a hospital ward’s nurse station. Nurses don’t just check on patients once — they have dashboards showing heart rate, oxygen levels, and alerts for anything abnormal. If a patient’s vitals dip, the alarm fires before it becomes a crisis.

Agent monitoring works the same way. You instrument your agents with “vital signs” — resolution rate, response time, error rate — and build dashboards that surface problems early. Without monitoring, you’re flying blind. An agent could silently fail 40 percent of the time and nobody would know until users start complaining.

Agent monitoring spans multiple layers: platform-level analytics in Copilot Studio, application-level telemetry in Application Insights, infrastructure-level metrics in Azure Monitor, and evaluation pipelines in Azure AI Foundry. Each layer answers different questions. Copilot Studio Analytics tells you what users are asking. Application Insights tells you how the agent is performing. Azure Monitor tells you whether the infrastructure is healthy. Foundry evaluation tells you how good the responses are.

The monitoring process is cyclical: define metrics, instrument agents, create dashboards, set alerts, review and tune. This is not a one-time setup — it runs continuously throughout the agent’s lifecycle. A well-monitored agent improves over time. An unmonitored agent degrades silently.

The Scenario

🤖 Jordan Reeves pulls Sam Nguyen into a meeting. CareFirst’s patient scheduling agent has been live for three weeks, but Jordan has no idea how it’s actually performing. Are patients getting the right appointments? How often does the agent escalate to a human? What’s the average response time?

Sam, CareFirst’s IT ops lead, needs to build a monitoring framework from scratch. Here’s how he approaches it.

Monitoring Tools Landscape

Not every tool does the same job. Here’s what each one is built for:

CapabilityCopilot Studio AnalyticsApplication InsightsAzure MonitorFoundry EvaluationCoE Toolkit
Primary FocusConversation analyticsApp-level telemetryInfrastructure healthResponse quality scoringGovernance and adoption
Built-in or CustomBuilt-in dashboardsCustom queries via KQLCustom alerts and dashboardsEvaluation pipelines you definePre-built Power BI reports
Best ForTopic hit rates, escalation rates, session volumeLatency, errors, dependency tracking, custom eventsCPU, memory, availability, costGroundedness, relevance, coherence of AI responsesOrg-wide agent inventory, usage trends, compliance
Skill LevelLow — point-and-clickMedium — requires KQL knowledgeMedium — alert rules and dashboardsMedium-High — evaluation flow designLow-Medium — Power BI consumption
When to UseDay-to-day conversation reviewDeep performance debuggingInfrastructure alertingPeriodic quality auditsExecutive reporting and governance

Key Agent Metrics

These are the vital signs Sam tracks for every agent:

MetricWhat It MeasuresHealthy RangeWhy It Matters
Resolution ratePercentage of conversations resolved without human escalationAbove 70 percentLow resolution means the agent isn’t solving problems
Escalation ratePercentage of conversations handed to a humanBelow 25 percentHigh escalation defeats the purpose of the agent
CSAT scoreCustomer satisfaction rating post-conversationAbove 4.0 out of 5Users may get answers but still have a bad experience
Average response timeTime from user message to agent replyUnder 3 secondsSlow responses frustrate users and increase abandonment
Fallback ratePercentage of messages the agent can’t match to a topicBelow 15 percentHigh fallback means gaps in the agent’s knowledge
Error ratePercentage of conversations that hit a system errorBelow 2 percentErrors break trust and require immediate investigation
💡

Exam Tip: The exam frames monitoring as a continuous process, not a one-time setup. If a question asks “what should you do AFTER deploying an agent,” monitoring is almost always part of the answer. Look for options that describe ongoing review cycles, not just initial dashboard creation.

The Monitoring Process

Sam follows a five-step cycle. Each step feeds the next:

Step 1: Define Metrics

Before touching any tool, Sam asks: “What does success look like for this agent?” For the scheduling agent, success means patients get the right appointment, quickly, without needing to call the front desk.

He defines concrete targets: resolution rate above 75 percent, average response time under 2.5 seconds, CSAT above 4.2.

Step 2: Instrument Agents

Sam enables Copilot Studio Analytics (on by default) and connects the agent to Application Insights by adding the Connection string in the agent’s Settings > Advanced. He also adds custom telemetry events — for example, a “SchedulingComplete” event that fires when the agent successfully books an appointment.

Step 3: Create Dashboards

Sam builds two dashboards:

  • Operational dashboard in Azure Monitor — real-time view of errors, response times, and availability. This is for the IT ops team.
  • Business dashboard in Power BI, pulling from Copilot Studio Analytics — resolution rates, top topics, escalation trends. This is for Jordan and the clinical leads.

Step 4: Set Alerts

Alerts are the early warning system. Sam creates:

  • Critical alert: Error rate exceeds 5 percent for 10 minutes — pages the on-call engineer
  • Warning alert: Escalation rate exceeds 30 percent over 1 hour — notifies Jordan via Teams
  • Informational alert: Weekly summary of CSAT trends — emailed to Dr. Obi

Step 5: Review and Tune

Every two weeks, Jordan and Sam review the dashboards together. They look for patterns: Did the fallback rate spike after a holiday? Are certain appointment types causing more escalations? This review feeds directly into the tuning process (covered in the next module).

💡

Deep Dive: Application Insights uses KQL (Kusto Query Language) for custom queries. A common exam scenario asks you to query conversation duration percentiles. The pattern is: customEvents | where name == "ConversationEnd" | summarize percentile(duration, 95) by bin(timestamp, 1h). You don’t need to memorise exact KQL syntax, but understanding that Application Insights enables this level of analysis is testable.

Flashcards

Question

What is the difference between resolution rate and escalation rate?

Click or press Enter to reveal answer

Answer

Resolution rate measures conversations the agent resolved WITHOUT human help. Escalation rate measures conversations handed TO a human. They are related but not inverses — a conversation can end unresolved without escalation (e.g., user abandons).

Click to flip back

Question

Which monitoring tool would you use to track the groundedness and coherence of agent responses?

Click or press Enter to reveal answer

Answer

Azure AI Foundry evaluation pipelines. Foundry lets you define evaluation criteria (groundedness, relevance, coherence) and run them against conversation logs. Copilot Studio Analytics shows volume and escalation, not response quality.

Click to flip back

Question

Why is the fallback rate metric important for agent health?

Click or press Enter to reveal answer

Answer

Fallback rate shows how often the agent cannot match a user message to any topic. A high fallback rate (above 15 percent) means there are gaps in the agent's knowledge base or topic coverage. It signals where you need to add or refine topics.

Click to flip back

Question

Name the five steps in the agent monitoring process cycle.

Click or press Enter to reveal answer

Answer

1. Define metrics and success criteria. 2. Instrument agents with telemetry. 3. Create operational and business dashboards. 4. Set alerts for critical thresholds. 5. Review and tune on a regular cadence. The cycle repeats continuously.

Click to flip back

Knowledge Check

Knowledge Check

Sam notices the scheduling agent's resolution rate dropped from 78 percent to 55 percent over the past week. Which monitoring action should he take FIRST?

Knowledge Check

Which combination of tools gives you BOTH real-time infrastructure alerting AND periodic AI response quality evaluation?

🎬 Video coming soon


Next up: Telemetry and Tuning — learn how to interpret telemetry data and tune agents based on what the metrics tell you.

← Previous

D365 AI Orchestration: Finance, SCM & Customer Experience

Next →

Telemetry Interpretation and Agent Tuning

Guided

I learn, I simplify, I share.

A Guide to Cloud YouTube Feedback

© 2026 Sutheesh. All rights reserved.

Guided is an independent study resource and is not affiliated with, endorsed by, or officially connected to Microsoft. Microsoft, Azure, and related trademarks are property of Microsoft Corporation. Always verify information against Microsoft Learn.