πŸ”’ Guided

Pre-launch preview. Authorised access only.

Incorrect code

Guided by A Guide to Cloud
Explore AB-900 AI-901
Guided AI-300 Domain 2
Domain 2 β€” Module 5 of 8 63%
10 of 25 overall

AI-300 Study Guide

Domain 1: Design and Implement an MLOps Infrastructure

  • ML Workspace: Your AI Control Room Free
  • Data, Environments & Components
  • Compute Targets: Choosing the Right Engine
  • Infrastructure as Code: Provisioning at Scale
  • Git & CI/CD for ML Projects

Domain 2: Implement Machine Learning Model Lifecycle and Operations

  • MLflow: Track Every Experiment Free
  • AutoML & Hyperparameter Tuning
  • Training Pipelines: Automate Everything
  • Distributed Training: Scale to Big Data
  • Model Registration & Versioning
  • Model Approval & Responsible AI Gates
  • Deploying Models: Endpoints in Production
  • Drift, Monitoring & Retraining

Domain 3: Design and Implement a GenAIOps Infrastructure

  • Foundry: Hubs, Projects & Platform Setup Free
  • Network Security & IaC for Foundry
  • Deploying Foundation Models
  • Model Versioning & Production Strategies
  • PromptOps: Design, Compare, Version & Ship

Domain 4: Implement Generative AI Quality Assurance and Observability

  • Evaluation: Datasets, Metrics & Quality Gates Free
  • Safety Evaluations & Custom Metrics
  • Monitoring GenAI in Production
  • Cost Tracking, Logging & Debugging

Domain 5: Optimize Generative AI Systems and Model Performance

  • RAG Optimization: Better Retrieval, Better Answers Free
  • Embeddings & Hybrid Search
  • Fine-Tuning: Methods, Data & Production

AI-300 Study Guide

Domain 1: Design and Implement an MLOps Infrastructure

  • ML Workspace: Your AI Control Room Free
  • Data, Environments & Components
  • Compute Targets: Choosing the Right Engine
  • Infrastructure as Code: Provisioning at Scale
  • Git & CI/CD for ML Projects

Domain 2: Implement Machine Learning Model Lifecycle and Operations

  • MLflow: Track Every Experiment Free
  • AutoML & Hyperparameter Tuning
  • Training Pipelines: Automate Everything
  • Distributed Training: Scale to Big Data
  • Model Registration & Versioning
  • Model Approval & Responsible AI Gates
  • Deploying Models: Endpoints in Production
  • Drift, Monitoring & Retraining

Domain 3: Design and Implement a GenAIOps Infrastructure

  • Foundry: Hubs, Projects & Platform Setup Free
  • Network Security & IaC for Foundry
  • Deploying Foundation Models
  • Model Versioning & Production Strategies
  • PromptOps: Design, Compare, Version & Ship

Domain 4: Implement Generative AI Quality Assurance and Observability

  • Evaluation: Datasets, Metrics & Quality Gates Free
  • Safety Evaluations & Custom Metrics
  • Monitoring GenAI in Production
  • Cost Tracking, Logging & Debugging

Domain 5: Optimize Generative AI Systems and Model Performance

  • RAG Optimization: Better Retrieval, Better Answers Free
  • Embeddings & Hybrid Search
  • Fine-Tuning: Methods, Data & Production
Domain 2: Implement Machine Learning Model Lifecycle and Operations Premium ⏱ ~12 min read

Model Registration & Versioning

A trained model is just a file until it's registered. Learn to package, register, version, and manage models through their lifecycle β€” from experiment to production to retirement.

From experiment to production artifact

β˜• Simple explanation

Think of a model like a recipe you’ve perfected.

You’ve tested 50 variations and finally nailed it. Now you need to: write it down properly (package), file it in a recipe book with a version number (register), and decide when to use it, update it, or retire it (lifecycle).

Without this process, your β€œbest model” is just a file on someone’s laptop. With it, every model is tracked, versioned, and ready for deployment.

Model registration bridges the gap between experimentation and deployment:

  • Packaging β€” standard MLflow model format with dependencies and metadata
  • Registration β€” versioned entry in the model registry with lineage tracking
  • Lifecycle management β€” staging, production, and archiving states
  • Feature retrieval spec β€” bundling feature definitions with the model so inference knows what inputs to expect

Registering an MLflow model

After a successful training run, register the model in the Azure ML model registry:

from azure.ai.ml.entities import Model
from azure.ai.ml.constants import AssetTypes

# Register from an MLflow run
model = Model(
    path="azureml://jobs/{run_id}/outputs/artifacts/churn-model",
    name="churn-predictor",
    version="3",
    type=AssetTypes.MLFLOW_MODEL,
    description="Churn prediction model β€” LightGBM, F1=0.961",
    tags={
        "algorithm": "lightgbm",
        "dataset_version": "v2",
        "f1_score": "0.961"
    }
)
registered_model = ml_client.models.create_or_update(model)

What’s happening:

  • Line 6: Points to the model artifact from a specific training run β€” full lineage
  • Line 9: MLFLOW_MODEL type β€” includes the model file, conda environment, and MLmodel metadata
  • Lines 11-14: Tags make the model searchable and auditable

Or register directly from MLflow:

import mlflow

# During training, log and register in one step
with mlflow.start_run():
    # ... train model ...
    mlflow.sklearn.log_model(model, "model",
        registered_model_name="churn-predictor")
πŸ’‘ Exam tip: MLflow model format

The MLflow model format is a directory containing:

  • MLmodel β€” metadata file listing flavors (sklearn, pytorch, etc.)
  • model.pkl (or equivalent) β€” the serialised model
  • conda.yaml β€” environment dependencies
  • requirements.txt β€” pip dependencies

This standard format means the model can be deployed to any MLflow-compatible platform. The exam favours MLflow model format over custom serialisation.

Feature retrieval specification

A feature retrieval spec bundles feature definitions with the model, ensuring that inference uses the correct features in the correct format:

from azure.ai.ml.entities import Model

model = Model(
    path="./model_output",
    name="churn-predictor",
    version="4",
    type=AssetTypes.MLFLOW_MODEL,
    properties={
        "feature_retrieval_spec": "feature_retrieval_spec.yaml"
    }
)

The spec file defines which features the model expects and where to retrieve them:

# feature_retrieval_spec.yaml
feature_store_uri: azureml://featurestores/churn-features
features:
  - name: customer_tenure_months
    source: customer_features:v2
  - name: monthly_charges
    source: billing_features:v1
  - name: support_tickets_30d
    source: support_features:v3
Scenario: Kai packages the churn model

Kai’s churn model needs 12 features to make predictions. Without a feature retrieval spec, the deployment team has to guess which features to send. With the spec bundled into the model:

  1. Training time: model is trained on features from the feature store
  2. Registration: feature spec is packaged with the model
  3. Deployment time: the endpoint reads the spec and automatically retrieves the right features
  4. Result: zero ambiguity about what the model needs

Priya (CTO): β€œNo more deployment bugs because someone sent the wrong features.”

Model lifecycle management

Models go through stages as they mature:

StageMeaningAction
None (default)Just registered, not validatedRun evaluation before promoting
StagingBeing tested, not serving production trafficDeploy to staging endpoint, run tests
ProductionActively serving predictionsDeploy to production endpoint
ArchivedRetired, kept for audit/complianceRemove from endpoints, retain in registry
# Update model stage
from azure.ai.ml.entities import Model

model = ml_client.models.get(name="churn-predictor", version="3")

# Archive an old version
ml_client.models.archive(name="churn-predictor", version="1")

# Restore if needed (for audit or rollback)
ml_client.models.restore(name="churn-predictor", version="1")
Scenario: Dr. Fatima's model governance at Meridian

Meridian Financial has strict model governance requirements. Dr. Fatima implements a lifecycle policy:

  1. Registration β€” every model must include: algorithm, dataset version, metrics, and feature spec
  2. Staging gate β€” model must pass responsible AI evaluation (Module 11) before staging
  3. Production gate β€” model must pass A/B test against current production model
  4. Archiving β€” retired models kept for 7 years (regulatory requirement)
  5. Version limit β€” max 10 active versions per model; older ones auto-archived

James Chen (CISO): β€œEvery model decision is auditable. That’s what regulators want to see.”

Key terms flashcards

Question

What is the MLflow model format?

Click or press Enter to reveal answer

Answer

A standardised directory containing: MLmodel (metadata), the serialised model file, conda.yaml (environment), and requirements.txt (pip deps). Deployable to any MLflow-compatible platform.

Click to flip back

Question

What is a feature retrieval specification?

Click or press Enter to reveal answer

Answer

A YAML file bundled with a model that defines which features the model needs and where to retrieve them. Ensures inference uses the correct features in the correct format.

Click to flip back

Question

What are the model lifecycle stages?

Click or press Enter to reveal answer

Answer

None (just registered) β†’ Staging (testing) β†’ Production (serving) β†’ Archived (retired, kept for audit). Models can be restored from archive for rollback.

Click to flip back

Knowledge check

Knowledge Check

Kai's deployment team keeps sending the wrong features to the churn model endpoint, causing prediction errors. What should Kai bundle with the registered model?

Knowledge Check

Dr. Fatima needs to retire an old model version but must keep it available for regulatory audits for 7 years. What should she do?

🎬 Video coming soon


Next up: Model Approval & Responsible AI Gates β€” ensuring models are fair, explainable, and safe before deployment.

← Previous

Distributed Training: Scale to Big Data

Next β†’

Model Approval & Responsible AI Gates

Guided

I learn, I simplify, I share.

A Guide to Cloud YouTube Feedback

© 2026 Sutheesh. All rights reserved.

Guided is an independent study resource and is not affiliated with, endorsed by, or officially connected to Microsoft. Microsoft, Azure, and related trademarks are property of Microsoft Corporation. Always verify information against Microsoft Learn.