πŸ”’ Guided

Pre-launch preview. Authorised access only.

Incorrect code

Guided by A Guide to Cloud
Explore AB-900 AI-901 aws-aif-c01
Guided AB-620 Domain 2
Domain 2 β€” Module 11 of 12 92%
21 of 28 overall

AB-620 Study Guide

Domain 1: Plan and Configure Agent Solutions

  • Getting Started: Copilot Studio for Developers Free
  • Planning Enterprise Integration and Reusable Components Free
  • Identity Strategy for Agents Free
  • Channels, Deployment and Audience Design Free
  • Responsible AI and Security Governance Free
  • Agent Flows: Build, Monitor and Handle Errors Free
  • Human-in-the-Loop Agent Flows Free
  • Topics, Tools and Variables Free
  • Advanced Responses: Custom Prompts and Generative Answers Free
  • API Calls, HTTP Requests and Adaptive Cards Free

Domain 2: Integrate and Extend Agents in Copilot Studio

  • Enterprise Knowledge Sources: The Big Picture
  • Copilot Connectors and Power Platform Connectors
  • Azure AI Search as a Knowledge Source
  • Adding Tools: Custom Connectors and REST APIs
  • MCP Tools: Model Context Protocol in Action
  • Computer Use: Agent-Driven UI Automation
  • Multi-Agent Solutions: Design and Agent Reuse
  • Integrating Foundry Agents
  • Fabric Data Agents: Analytics Meets AI
  • A2A Protocol: Cross-Platform Agent Collaboration
  • Grounded Answers: Azure AI Search with Foundry
  • Foundry Model Catalog and Application Insights

Domain 3: Test and Manage Agents

  • Test Sets & Evaluation Methods
  • Reviewing Results & Tuning Performance
  • Solutions & Environment Variables
  • Power Platform Pipelines for Agent ALM
  • Agent Lifecycle: From Dev to Production
  • Exam Prep: Diagnostic Review

AB-620 Study Guide

Domain 1: Plan and Configure Agent Solutions

  • Getting Started: Copilot Studio for Developers Free
  • Planning Enterprise Integration and Reusable Components Free
  • Identity Strategy for Agents Free
  • Channels, Deployment and Audience Design Free
  • Responsible AI and Security Governance Free
  • Agent Flows: Build, Monitor and Handle Errors Free
  • Human-in-the-Loop Agent Flows Free
  • Topics, Tools and Variables Free
  • Advanced Responses: Custom Prompts and Generative Answers Free
  • API Calls, HTTP Requests and Adaptive Cards Free

Domain 2: Integrate and Extend Agents in Copilot Studio

  • Enterprise Knowledge Sources: The Big Picture
  • Copilot Connectors and Power Platform Connectors
  • Azure AI Search as a Knowledge Source
  • Adding Tools: Custom Connectors and REST APIs
  • MCP Tools: Model Context Protocol in Action
  • Computer Use: Agent-Driven UI Automation
  • Multi-Agent Solutions: Design and Agent Reuse
  • Integrating Foundry Agents
  • Fabric Data Agents: Analytics Meets AI
  • A2A Protocol: Cross-Platform Agent Collaboration
  • Grounded Answers: Azure AI Search with Foundry
  • Foundry Model Catalog and Application Insights

Domain 3: Test and Manage Agents

  • Test Sets & Evaluation Methods
  • Reviewing Results & Tuning Performance
  • Solutions & Environment Variables
  • Power Platform Pipelines for Agent ALM
  • Agent Lifecycle: From Dev to Production
  • Exam Prep: Diagnostic Review
Domain 2: Integrate and Extend Agents in Copilot Studio Premium ⏱ ~14 min read

Grounded Answers: Azure AI Search with Foundry

Configure a production-grade RAG pipeline using Azure AI Search and Foundry to ground your agent's generative answers in trusted enterprise documents.

What is grounding β€” and why does it matter?

β˜• Simple explanation

Imagine an AI agent answering medical questions from memory alone.

Sometimes it gets things right. Sometimes it confidently makes things up β€” and in healthcare, that is dangerous. Grounding is like giving the agent a textbook and saying β€œonly answer from this book.” Before the agent responds, it searches your trusted documents, finds the relevant passages, and uses those to craft its answer. If the textbook does not cover the topic, the agent says β€œI do not know” instead of guessing.

Azure AI Search is the search engine that indexes your documents. Foundry is the AI brain that reads the search results and generates an answer. Together, they create a RAG pipeline β€” Retrieval-Augmented Generation β€” the gold standard for trustworthy AI answers.

Grounding means constraining an LLM’s responses to information retrieved from authoritative sources rather than relying solely on the model’s training data. This dramatically reduces hallucination and enables citations.

A RAG (Retrieval-Augmented Generation) pipeline has four stages: Index (documents are chunked, embedded, and stored in a search index), Retrieve (user query is used to find relevant chunks), Augment (retrieved chunks are injected into the LLM prompt as context), and Generate (the LLM produces an answer grounded in the retrieved context).

In the Copilot Studio + Foundry pattern, Azure AI Search handles indexing and retrieval, while a Foundry model handles augmentation and generation. This differs from Module 13’s basic generative answers (which use Copilot Studio’s built-in connection) β€” here you get full control over the search index, chunking strategy, embedding model, and generation model.

The RAG pipeline explained

Understanding the four stages is critical for the exam. Each stage has configuration decisions that affect answer quality.

StageWhat happensKey decisions
1. IndexDocuments are chunked into segments, converted to vector embeddings, and stored in Azure AI SearchChunk size (smaller = precise, larger = more context), embedding model choice, which documents to include
2. RetrieveUser’s question is converted to an embedding and matched against the index using hybrid search (vector + keyword)Number of results to return (top-k), search mode (vector, keyword, or hybrid), filters (metadata, date, category)
3. AugmentRetrieved chunks are injected into the LLM prompt as context: β€œAnswer the user’s question using ONLY the following sources…”Prompt template design, how many chunks to include, whether to include metadata
4. GenerateThe Foundry model reads the augmented prompt and generates a grounded answer with citationsModel choice (GPT-4o for accuracy, GPT-4o mini for speed/cost), temperature setting, citation format
ℹ️ Why RAG instead of fine-tuning?

Fine-tuning changes the model itself β€” you train it on your data. RAG keeps the model unchanged and feeds it relevant documents at query time. For enterprise knowledge that changes frequently (policies, procedures, product docs), RAG is almost always better because:

  • No retraining needed when documents update β€” just re-index
  • Citations are traceable β€” the agent can point to the exact document and passage
  • Security is enforced at the index level β€” you control which documents each user can access
  • Cost is lower β€” indexing is cheaper than fine-tuning runs

The exam assumes RAG for enterprise grounding scenarios. Fine-tuning appears in Foundry model catalog contexts (Module 22).

How this differs from basic generative answers

You already learned about generative answers in the enterprise knowledge module β€” SharePoint, Dataverse, website knowledge sources. The Foundry + Azure AI Search pattern is the production-grade version.

Basic vs Foundry-grounded generative answers
FeatureControl levelKnowledge sourcesBest for
Basic generative answersLow β€” Copilot Studio manages everything (chunking, search, generation) behind the scenesSharePoint, Dataverse, public websites, uploaded files β€” configured directly in Copilot StudioQuick setup, internal docs, FAQ-style answers where default behavior is good enough
Azure AI Search + FoundryFull β€” you control the index schema, chunking, embedding model, search strategy, generation model, and prompt templateAny documents indexed in Azure AI Search β€” PDFs, databases, APIs, custom data pipelinesProduction workloads needing precision, custom ranking, citations, compliance, or domain-specific models

Configuration steps

Setting up the Foundry + Azure AI Search pipeline involves both Azure and Copilot Studio:

Azure side (your AI engineer sets this up):

  1. Create an Azure AI Search resource β€” choose tier based on data volume and query load
  2. Create a search index β€” define the schema (fields, types, searchable/filterable attributes)
  3. Index your documents β€” use an indexer to chunk documents, generate embeddings, and populate the index
  4. Test search quality β€” run sample queries to verify relevant results are returned
  5. Create a Foundry project β€” connect it to the Azure AI Search index

Copilot Studio side (the agent developer connects it):

  1. In the agent, navigate to Knowledge then Add knowledge then Azure AI Search with Foundry
  2. Provide the Foundry project details β€” endpoint, API key, and index name
  3. Configure the generative answers node β€” set it up in the topic where grounded answers are needed
  4. Set citation behavior β€” choose whether to show source document names, URLs, or passage excerpts
  5. Test with real questions β€” verify answers are grounded in the indexed documents, not hallucinated
πŸ’‘ Key concepts to remember for the exam

The exam tests configuration, not Azure infrastructure setup. Focus on:

  • RAG = Retrieval-Augmented Generation β€” the four-stage pipeline (index, retrieve, augment, generate)
  • Grounding = constraining answers to retrieved sources β€” reduces hallucination, enables citations
  • Azure AI Search + Foundry gives full control β€” basic generative answers use defaults you cannot customize
  • Citations are a feature of grounding β€” the agent can reference the exact document that informed its answer
  • This is different from Module 13 β€” Module 13 covers basic knowledge source connections; this module covers the production-grade Foundry pipeline
Scenario: Lena configures clinical answers grounded in research papers

Lena’s healthcare firm has a library of 12,000 peer-reviewed medical research papers and 3,000 internal clinical guidelines. Doctors need an agent that answers questions like β€œWhat are the current treatment protocols for Type 2 diabetes in patients over 65?” β€” and every answer must cite the specific paper or guideline it draws from.

Basic generative answers will not work here: Lena needs control over chunking (medical papers have complex structures), search ranking (recent papers should rank higher), and citations (doctors must verify every recommendation).

She builds the pipeline: Azure AI Search indexes all 15,000 documents with metadata (publication date, specialty, evidence level). The chunking strategy splits papers by section (abstract, methods, results, discussion) so retrieval is precise. She selects GPT-4o in Foundry for generation β€” it handles medical terminology better than smaller models.

The Copilot Studio developer connects this pipeline as a knowledge source in the hospital agent. When a doctor asks about diabetes treatment, the agent searches the index, retrieves the five most relevant paper sections, and generates an answer with inline citations: β€œAccording to Smith et al. (2024) in the Journal of Endocrinology, current first-line treatment for Type 2 diabetes in elderly patients is…”

Every answer is traceable. Every citation is verifiable. This is grounding in action.

Question

What does RAG stand for and what are its four stages?

Click or press Enter to reveal answer

Answer

Retrieval-Augmented Generation. Stages: 1) Index β€” chunk and embed documents into a search index. 2) Retrieve β€” find relevant chunks for the user's query. 3) Augment β€” inject retrieved chunks into the LLM prompt. 4) Generate β€” LLM produces a grounded answer from the context.

Click to flip back

Question

What is the difference between basic generative answers and Azure AI Search + Foundry grounding?

Click or press Enter to reveal answer

Answer

Basic: Copilot Studio manages everything with defaults β€” quick setup, limited control. Foundry: you control the search index, chunking, embedding model, generation model, and prompt β€” full production-grade customisation with citations.

Click to flip back

Question

What does 'grounding' mean in the context of AI agents?

Click or press Enter to reveal answer

Answer

Constraining an LLM's responses to information retrieved from authoritative sources, rather than relying on the model's training data. This reduces hallucination and enables traceable citations.

Click to flip back

Knowledge Check

Lena needs her medical agent to cite specific research papers when answering clinical questions. Which approach should she use?

Knowledge Check

In a RAG pipeline, what happens during the 'Augment' stage?

Knowledge Check

Why is RAG preferred over fine-tuning for enterprise knowledge grounding?

🎬 Video coming soon

Grounded Answers: Azure AI Search with Foundry

← Previous

A2A Protocol: Cross-Platform Agent Collaboration

Next β†’

Foundry Model Catalog and Application Insights

Guided

I learn, I simplify, I share.

A Guide to Cloud YouTube Feedback

© 2026 Sutheesh. All rights reserved.

Guided is an independent study resource and is not affiliated with, endorsed by, or officially connected to Microsoft. Microsoft, Azure, and related trademarks are property of Microsoft Corporation. Always verify information against Microsoft Learn.