🔒 Guided

Pre-launch preview. Authorised access only.

Incorrect code

Guided by A Guide to Cloud
Explore AB-900 AI-901 aws-aif-c01
Guided AB-620 Domain 2
Domain 2 — Module 12 of 12 100%
22 of 28 overall

AB-620 Study Guide

Domain 1: Plan and Configure Agent Solutions

  • Getting Started: Copilot Studio for Developers Free
  • Planning Enterprise Integration and Reusable Components Free
  • Identity Strategy for Agents Free
  • Channels, Deployment and Audience Design Free
  • Responsible AI and Security Governance Free
  • Agent Flows: Build, Monitor and Handle Errors Free
  • Human-in-the-Loop Agent Flows Free
  • Topics, Tools and Variables Free
  • Advanced Responses: Custom Prompts and Generative Answers Free
  • API Calls, HTTP Requests and Adaptive Cards Free

Domain 2: Integrate and Extend Agents in Copilot Studio

  • Enterprise Knowledge Sources: The Big Picture
  • Copilot Connectors and Power Platform Connectors
  • Azure AI Search as a Knowledge Source
  • Adding Tools: Custom Connectors and REST APIs
  • MCP Tools: Model Context Protocol in Action
  • Computer Use: Agent-Driven UI Automation
  • Multi-Agent Solutions: Design and Agent Reuse
  • Integrating Foundry Agents
  • Fabric Data Agents: Analytics Meets AI
  • A2A Protocol: Cross-Platform Agent Collaboration
  • Grounded Answers: Azure AI Search with Foundry
  • Foundry Model Catalog and Application Insights

Domain 3: Test and Manage Agents

  • Test Sets & Evaluation Methods
  • Reviewing Results & Tuning Performance
  • Solutions & Environment Variables
  • Power Platform Pipelines for Agent ALM
  • Agent Lifecycle: From Dev to Production
  • Exam Prep: Diagnostic Review

AB-620 Study Guide

Domain 1: Plan and Configure Agent Solutions

  • Getting Started: Copilot Studio for Developers Free
  • Planning Enterprise Integration and Reusable Components Free
  • Identity Strategy for Agents Free
  • Channels, Deployment and Audience Design Free
  • Responsible AI and Security Governance Free
  • Agent Flows: Build, Monitor and Handle Errors Free
  • Human-in-the-Loop Agent Flows Free
  • Topics, Tools and Variables Free
  • Advanced Responses: Custom Prompts and Generative Answers Free
  • API Calls, HTTP Requests and Adaptive Cards Free

Domain 2: Integrate and Extend Agents in Copilot Studio

  • Enterprise Knowledge Sources: The Big Picture
  • Copilot Connectors and Power Platform Connectors
  • Azure AI Search as a Knowledge Source
  • Adding Tools: Custom Connectors and REST APIs
  • MCP Tools: Model Context Protocol in Action
  • Computer Use: Agent-Driven UI Automation
  • Multi-Agent Solutions: Design and Agent Reuse
  • Integrating Foundry Agents
  • Fabric Data Agents: Analytics Meets AI
  • A2A Protocol: Cross-Platform Agent Collaboration
  • Grounded Answers: Azure AI Search with Foundry
  • Foundry Model Catalog and Application Insights

Domain 3: Test and Manage Agents

  • Test Sets & Evaluation Methods
  • Reviewing Results & Tuning Performance
  • Solutions & Environment Variables
  • Power Platform Pipelines for Agent ALM
  • Agent Lifecycle: From Dev to Production
  • Exam Prep: Diagnostic Review
Domain 2: Integrate and Extend Agents in Copilot Studio Premium ⏱ ~14 min read

Foundry Model Catalog and Application Insights

Choose the right AI model from the Foundry model catalog for custom prompts, and monitor your agents with Application Insights telemetry.

Part 1: The Foundry Model Catalog

☕ Simple explanation

Think of the model catalog like a car dealership.

You would not buy a sports car to deliver furniture, and you would not buy a delivery van for a race. AI models are the same — some are fast and cheap (great for simple FAQs), others are powerful and expensive (needed for complex reasoning). The Foundry model catalog is your dealership: hundreds of models from Microsoft, OpenAI, Meta, and others, each with different strengths.

For Copilot Studio, the model catalog matters when you create custom prompts — special AI instructions that run inside your agent’s topics. Instead of using the default model, you pick the model that best fits each task.

Application Insights is your agent’s dashboard — it tracks every conversation, measures response times, catches errors, and shows you how your agent performs in the real world.

The Foundry model catalog provides access to hundreds of AI models spanning multiple providers: OpenAI (GPT-4o, GPT-4o mini), Microsoft (Phi family), Meta (Llama), Mistral, Cohere, and more. Each model varies in cost, latency, capability, and context window size.

In Copilot Studio, the model catalog is relevant when configuring custom prompts — prompt nodes within topics that call a specific Foundry model instead of the agent’s default model. This enables model-per-task optimization: use GPT-4o for complex medical reasoning, GPT-4o mini for simple classification, and Phi for high-volume low-cost tasks.

Application Insights is Azure’s application performance monitoring (APM) service. When connected to a Copilot Studio agent, it captures conversation telemetry — session counts, topic completion rates, resolution rates, error rates, and custom events. This data powers KQL (Kusto Query Language) queries and dashboards for operational monitoring.

Choosing the right model

Not all AI models are equal. The exam expects you to match model characteristics to use cases.

Foundry model catalog — key models compared
FeatureCostSpeedCapabilityBest for
GPT-4oHigher — premium per-token pricingModerate — 10-30 seconds for complex reasoningHighest — complex reasoning, nuanced understanding, multi-step analysis, medical/legal/financial domainsHigh-stakes tasks: clinical decision support, contract analysis, complex troubleshooting
GPT-4o miniLow — fraction of GPT-4o costFast — typically under 5 secondsGood — handles most business tasks well, strong at summarisation and classificationHigh-volume tasks: FAQ answers, ticket classification, simple summarisation, routing logic
Phi (small language model)Lowest — designed for cost efficiency at scaleFastest — sub-second for simple tasksModerate — strong for structured tasks, weaker on open-ended reasoningEdge deployment, high-throughput classification, cost-sensitive scenarios with thousands of daily calls
Llama (Meta)Varies by size — competitive with GPT-4o miniVaries — depends on model size (8B, 70B, 405B)Strong — open-source, good at general reasoning, code generationTeams preferring open-source models, specific compliance requirements, code-heavy tasks

Custom prompts with model selection

A custom prompt in Copilot Studio is a prompt node within a topic that sends a specific instruction to an AI model and returns the result. By connecting to the Foundry model catalog, you can choose which model processes each prompt.

Configuration steps:

  1. Connect your Copilot Studio environment to Foundry — this enables model catalog access
  2. In a topic, add a Prompt node (also called a “Create prompt” or “AI Builder prompt” action)
  3. Select the Foundry model — choose from the catalog based on your task requirements
  4. Write the prompt instruction — what the model should do with the input (e.g., “Classify this ticket as urgent, normal, or low priority”)
  5. Map input variables — pass conversation variables into the prompt
  6. Map output variables — capture the model’s response for use in subsequent nodes
  7. Test with sample inputs — verify the model produces expected outputs
ℹ️ Why not just use the default model for everything?

Custom prompts with specific models give three advantages: cost optimization (GPT-4o mini for simple tasks, GPT-4o only for complex reasoning — can cut AI costs 60-80%), task-specific accuracy (some models excel at certain tasks), and compliance (choose models deployed in specific Azure regions for data residency requirements).

Part 2: Monitoring with Application Insights

Building an agent is half the job. The other half is knowing whether it actually works in production. Application Insights provides the observability layer.

Connecting Application Insights to your agent:

  1. Create an Application Insights resource in Azure (or use an existing one)
  2. In Copilot Studio, go to Settings then Agent settings then Application Insights
  3. Paste the connection string from your Application Insights resource
  4. Save and publish — telemetry starts flowing within minutes

Key telemetry captured:

MetricWhat it tells youWhy it matters
Session countHow many conversations happen per day/weekAdoption tracking — is the agent being used?
Topic completion ratePercentage of topic starts that reach the end nodeQuality signal — incomplete topics suggest confusion or errors
Resolution ratePercentage of sessions resolved without human escalationEffectiveness — the agent’s core success metric
Escalation rateHow often conversations transfer to a humanCapacity planning — high escalation means the agent needs improvement
Average response timeHow long the agent takes to respondUser experience — slow responses increase abandonment
Error rateFailed connector calls, timeout errors, unhandled exceptionsReliability — errors need immediate investigation
ℹ️ KQL query examples for agent monitoring

Application Insights data is queried using KQL (Kusto Query Language). Common queries include sessions per day (summarize dcount(session_Id) by bin(timestamp, 1d)), top escalated topics, and average response latency.

You do not need to memorise KQL syntax for the exam — but knowing that Application Insights enables query-driven monitoring is testable.

Scenario: Lena picks models and sets up monitoring

Lena’s hospital agent handles two workloads: clinical decision support (complex medical questions requiring accuracy) and general FAQ (parking, cafeteria, IT password resets — speed and cost matter more).

For clinical custom prompts she selects GPT-4o — its reasoning accuracy on medical terminology is worth the premium. For FAQ prompts she picks GPT-4o mini — simple Q&A at a fraction of the cost, under 3 seconds.

She connects Application Insights and builds dashboards tracking clinical accuracy, daily usage, and cost per model. After the first week, data reveals 40% of “clinical” queries are simple medication lookups. She adds a classification prompt (GPT-4o mini) that routes simple lookups to the cheaper model. Cost drops 35% with no accuracy impact. Data-driven model optimization in action.

💡 Exam tip: model selection is about matching cost and capability to the task

The exam will describe scenarios and ask which model to choose. The decision framework is simple:

  • Complex reasoning, high stakes → GPT-4o (or the most capable model available)
  • Simple tasks, high volume → GPT-4o mini (good balance of cost and capability)
  • Maximum cost efficiency, structured tasks → Phi (smallest, cheapest, fastest)
  • Open-source requirement → Llama

If the scenario mentions “thousands of daily requests” and “simple classification,” the answer is almost always GPT-4o mini or Phi — not GPT-4o.

Question

What is the Foundry model catalog?

Click or press Enter to reveal answer

Answer

A library of hundreds of AI models from multiple providers (OpenAI, Microsoft, Meta, Mistral, etc.) available in Microsoft Foundry. Copilot Studio developers use it to select specific models for custom prompts based on cost, speed, and capability requirements.

Click to flip back

Question

When should you use GPT-4o vs GPT-4o mini in custom prompts?

Click or press Enter to reveal answer

Answer

GPT-4o: complex reasoning, high-stakes tasks (medical, legal, financial) where accuracy is critical. GPT-4o mini: high-volume, simpler tasks (FAQ, classification, summarisation) where cost and speed matter more than reasoning depth.

Click to flip back

Question

How do you connect Application Insights to a Copilot Studio agent?

Click or press Enter to reveal answer

Answer

Create an Application Insights resource in Azure → In Copilot Studio go to Settings → Agent settings → Application Insights → paste the connection string → Save and publish. Telemetry starts flowing within minutes.

Click to flip back

Question

Name four key metrics captured by Application Insights for Copilot Studio agents.

Click or press Enter to reveal answer

Answer

1) Session count (adoption). 2) Topic completion rate (quality). 3) Resolution rate (effectiveness). 4) Escalation rate (agent capability gaps). Also: response time and error rate.

Click to flip back

Question

What query language is used to analyse Application Insights data?

Click or press Enter to reveal answer

Answer

KQL (Kusto Query Language). It enables queries against agent telemetry — session counts, topic performance, error rates, response latency, and custom events.

Click to flip back

Knowledge Check

Lena's agent handles thousands of simple FAQ questions daily and a few dozen complex clinical queries. How should she configure model selection?

Knowledge Check

After connecting Application Insights, which metric best indicates that the agent is failing to help users?

Knowledge Check

What is the primary benefit of using the Foundry model catalog with custom prompts instead of the default Copilot Studio model?

🎬 Video coming soon

Foundry Model Catalog and Application Insights

← Previous

Grounded Answers: Azure AI Search with Foundry

Next →

Test Sets & Evaluation Methods

Guided

I learn, I simplify, I share.

A Guide to Cloud YouTube Feedback

© 2026 Sutheesh. All rights reserved.

Guided is an independent study resource and is not affiliated with, endorsed by, or officially connected to Microsoft. Microsoft, Azure, and related trademarks are property of Microsoft Corporation. Always verify information against Microsoft Learn.