RAG and Grounding: Making AI Use YOUR Data
Out-of-the-box AI only knows what it was trained on. Grounding connects it to your organisation's data — and RAG is the most common way to do it. Learn when and how to use both.
Why does AI need your data?
Think of a large language model like a well-educated new hire on their first day.
They know a lot about the world — business, writing, analysis — but they know nothing about YOUR company. Your policies, your products, your customers, your internal processes. If you ask them a company-specific question, they’ll either say “I don’t know” or make something up that sounds right.
Grounding is giving that new hire a cheat sheet — your company’s actual documents, data, and knowledge — so their answers are based on real information, not guesses.
RAG (Retrieval-Augmented Generation) is the most common way to hand over that cheat sheet: find the right documents first, then let AI answer based on what it found.
How RAG works: Three steps
RAG follows a simple pattern every time a user asks a question:
| Step | What happens | Example |
|---|---|---|
| 1. Retrieve | The system searches your knowledge base for relevant documents | User asks “What’s our return policy?” — the system finds the returns policy document, the FAQ page, and a recent policy update email |
| 2. Augment | The retrieved documents are added to the prompt as context | The AI now has the actual policy text in front of it — not just its training data |
| 3. Generate | The model produces an answer grounded in the retrieved content | ”Our return policy allows returns within 30 days with receipt. As of March 2026, we also accept digital receipts.” |
Without RAG, the model would have to guess your return policy — and it would probably fabricate something plausible but wrong.
Exam tip: RAG vs fine-tuning
The exam may ask you to choose between RAG and fine-tuning. Key distinctions:
- RAG retrieves data at query time. Data stays current. No model retraining needed. Respects existing access controls.
- Fine-tuning bakes knowledge into the model itself. Requires retraining when data changes. Better for teaching the model a specific style or behaviour.
For most business scenarios — answering questions over company data, policy lookups, customer support — RAG is the right answer. Fine-tuning is for specialised behaviour, not data retrieval.
Grounding in the Microsoft ecosystem
Microsoft offers two primary grounding approaches:
| Feature | How it works | Data sources | Best for |
|---|---|---|---|
| Microsoft 365 Copilot (Graph grounding) | Automatically grounded in your M365 data via Microsoft Graph | Emails, Teams chats, SharePoint documents, OneDrive files, calendar events | Knowledge workers who need AI that understands their work context — meetings, emails, documents |
| Azure AI Search + Azure OpenAI (Custom RAG) | You build a custom retrieval pipeline over your own data sources | Databases, APIs, file shares, CRM systems, custom knowledge bases — any data you index | Custom applications, customer-facing chatbots, or scenarios where M365 data isn't sufficient |
How Copilot grounding works behind the scenes
When you ask Microsoft 365 Copilot a question, here’s what happens:
- Your prompt goes to the Copilot orchestrator
- The orchestrator queries Microsoft Graph for relevant content you have access to — emails, files, chats, meetings
- The relevant content is injected into the prompt as grounding data
- The LLM generates a response based on your actual data
- The response goes through responsible AI filters before reaching you
Critically, Copilot only retrieves data the user already has permission to access. If you can’t see a document in SharePoint, Copilot can’t use it either.
Business requirements for grounding
Before deploying a grounded AI solution, leaders need to evaluate five key areas:
| Requirement | What to assess | Risk if ignored |
|---|---|---|
| Data quality | Is your data accurate, complete, and well-organised? | AI will ground on bad data and produce confidently wrong answers |
| Access control | Are document permissions correctly configured? | AI could surface confidential data to unauthorised users |
| Data freshness | How current is the indexed data? | AI answers based on outdated information — policy changes missed |
| Relevance | Does the retrieval system find the right documents? | AI grounds on irrelevant content, producing unhelpful or misleading responses |
| Coverage | Is all necessary knowledge indexed and searchable? | AI says “I don’t know” for questions it should be able to answer |
The hidden risk: Oversharing through AI
This is a common exam topic. When organisations deploy Copilot without reviewing their SharePoint permissions, they often discover that employees can see documents they shouldn’t — salary data, board minutes, HR cases.
Before Copilot, nobody noticed because nobody searched for that content. Now Copilot proactively surfaces it. The fix isn’t to restrict Copilot — it’s to fix the underlying permissions. AI doesn’t create the access problem. It reveals it.
Real-world scenario: Ravi builds grounded customer support
🏗️ Ravi, CTO of TechVantage Solutions, needs a customer support chatbot that answers questions about their software products. Out-of-the-box ChatGPT knows nothing about TechVantage’s products.
Ravi’s team builds a RAG solution:
- Index all product documentation, release notes, and known issues into Azure AI Search
- Connect Azure AI Search to Azure OpenAI Service as a grounding source
- Configure the system prompt: “You are a TechVantage support assistant. Only answer from the provided documents. If you don’t have the information, say so.”
- Deploy the chatbot on the customer portal
Results after 8 weeks:
- 85% of customer questions answered without human escalation (up from 30% with their old rule-based chatbot)
- Fabrication rate below 3% (the “say I don’t know” instruction works)
- Customer satisfaction up 22 points
- Support team focuses on complex issues instead of answering the same questions repeatedly
Why Ravi chose RAG over fine-tuning
Ravi considered fine-tuning but chose RAG because:
- Product docs change weekly — fine-tuning would require constant retraining
- He needs the bot to cite specific documents as sources
- RAG lets him update the knowledge base without touching the model
- Access controls ensure customers only see documentation for products they’ve purchased
Fine-tuning would have been appropriate if the goal was to change the model’s communication style or teach it a domain-specific language.
Key flashcards
Knowledge check
Ravi's support chatbot answers product questions using indexed documentation. When a customer asks about a feature, the system finds relevant docs, includes them in the prompt, and the AI generates an answer. What is this architecture called?
Dr. Patel is called in after a Copilot deployment at a financial services firm. Employees have discovered they can ask Copilot about executive salary data stored in a SharePoint site. What is the root cause?
🎬 Video coming soon
Next up: Data Quality: The Make-or-Break Factor for AI — why clean, representative data is the foundation of every successful AI deployment.