AI-900 Decision Framework
A decision and keyword atlas designed to support the AI-900 Readiness Assessment
Use this framework when unsure during the assessment. Focus on Keywords, Ask Yourself, and Traps.
β¨ Generative AI
Azure OpenAI
- Chat
- Text generation
- Summarization
- Code generation
Is the system creating new content?
Embeddings
- Numerical representations of meaning
- Semantic similarity
- Clustering & vector search
Are we comparing meaning, not keywords?
Tokens & Tokenization
- How a model reads text
- Text β Tokens β IDs β Processing
- Determines input limits and cost
Is the question about limits, length, or input size?
Prompt Engineering
- Writing better instructions for generative models
- AIβ900 scope: concept only (no syntax, no parameters)
Is it about improving how you ask the model?
β Does NOT apply to AutoML, Vision, or Language.
π Search & Knowledge Mining
Azure AI Search
- Search documents
- Index data
- Semantic search
Are users searching existing content?
Knowledge Mining (Hybrid)
- Search + AI enrichment
- Extract insights from large document collections
- Combines Azure AI Search with Language / Vision
Do we need to discover patterns across many documents?
π€ Conversational AI
Azure Bot Service
- Chatbots
- FAQ bots
Is the user talking to a system conversationally?
π‘οΈ Content Safety
Azure AI Content Safety
- Detect harmful content
- Text and image moderation
Is the goal to block or flag unsafe content?
π§© Azure AI Personalizer
RealβTime Personalization
- Chooses the best action per user in real time
- Learns from user feedback
- Optimizes decisions over time (continuously improves)
Is the system choosing one best option per user, not just predicting preferences?
β Personalizer β static rules
β Personalizer = adaptive decision making
Typical Use Cases
- Personalized product banners
- News or content ranking
- "Next best action" decisions
- Adaptive UI choices
ποΈ Azure AI Vision
Image Classification
- What is in the image (label)
- Single category output
Do we only need to know what it is?
Object Detection
- What + Where (bounding boxes)
- Can detect condition/state if visually inferable
- Examples: helmet worn, damage, open/closed
Do we need to know what AND where?
Face Detection
- Detect faces + location
- Age, emotion, accessories
| Task | Output |
|---|---|
| Image Classification | What is in the image (label) |
| Object Detection | What + where (bounding box) |
| Face Detection | Faces + location |
π OCR vs Document Intelligence
OCR
- Extract raw text from images
- Printed and handwritten
- Photos, screenshots, scanned pages
Is the goal to get raw text out of an image?
Document Intelligence
- Extract structured information from documents
- Invoices, receipts, forms, contracts
- Keyβvalue pairs, tables, fields
Do we need structured fields (not just raw text)?
| Service | Input | Output |
|---|---|---|
| OCR | Image with text | Raw unstructured text |
| Document Intelligence | Forms, invoices, receipts | Structured keyβvalue pairs, tables, fields |
π¬ Azure AI Language
Named Entity Recognition (NER)
- Extracts named things from text
- People, organizations, locations
- Dates, money, quantities
Are we pulling structured facts from text?
Sentiment Analysis
- Positive / Negative / Neutral
- Opinion mining
Are we measuring how people feel?
Key Phrases & Text Classification
- Extract important phrases from text
- Classify text into categories
Do we need to identify important topics or categorize text?
ποΈ Azure AI Speech
What it CAN do
- β SpeechβtoβText (transcription)
- β TextβtoβSpeech (voice output)
- β Speech translation
- β Speaker recognition
Is the input or output audio?
What it CANNOT do
- β Understand image context
- β Extract entities from text (β Language)
- β Perform sentiment analysis (β Language)
- β Replace a chatbot alone
Speech = audio processing Β· Language = text understanding
π Speech vs Language vs Bots β Separator
| Service | Handles | Input/Output |
|---|---|---|
| Azure AI Speech | Transcription, TTS, translation, speaker ID | Audio |
| Azure AI Language | NER, sentiment, key phrases, classification | Text |
| Conversational AI (Bots) | Dialogs, conversation flow, user interaction | Conversation management |
π Text Translation
Azure AI Translator
- Converts text from one language to another
- Preserves meaning across languages
Is the goal to convert text between languages, not create or analyze content?
β Translation β speechβtoβtext (audio)
β Translation does NOT summarize or analyze sentiment
Typical Use Cases
- Multilingual websites
- Customer support translation
- Global content distribution
π Entity Linking
Extension of NER
- NER identifies: "Microsoft" β Organization
- Entity Linking connects it to the known entity in a database
- Disambiguates: "Apple" the company vs fruit
Do we need to connect a mention to a known entity, not just label it?
Why It Matters
- Improves search accuracy
- Helps build knowledge graphs
- Avoids duplicates across mentions
β No implementation details required for AIβ900
π Core ML Patterns
Classification
- Predict a label / category
- Yes/No, type, class
Is the output a category or label?
Regression
- Predict a number
- Price, amount, score, quantity
Is the output a numerical value?
Clustering
- Group similar items
- No labels (unsupervised)
Are there no predefined categories?
Anomaly Detection
- Detect unusual behavior
- Fraud, spikes, failures
Is the goal to find what looks different from normal?
Recommendation
- Product suggestions
- User preferences
Is the system predicting what someone might like?
| Technique | Output | Use when |
|---|---|---|
| Classification | Label / category | Yes/No, type, class |
| Regression | Number | Price, amount, score |
| Clustering | Groups | No labels, similarity |
π§ͺ Azure Machine Learning
Automated ML (AutoML)
- Automatically tries multiple algorithms
- Optimizes model selection & parameters
- Less control, faster experimentation
Do you want the best model quickly without designing pipelines?
ML Designer
- Visual dragβandβdrop pipeline
- Explicit control over each step
- Full transparency
Do you want to define how the model is built?
Model Versioning
- Model registration & versioning
- Experiment tracking
- Deployment history
Do you need to track multiple versions of a model?
β Azure OpenAI does NOT version your custom ML models.
Inference
- Using a trained model to make predictions
- The "production" phase of ML
Is the model already trained and now being used?
| Feature | AutoML | ML Designer |
|---|---|---|
| Algorithm selection | Automatic | Manual |
| Speed | Faster | Slower |
| Control | Low | High |
| Visual pipeline | No | Yes |
| Phase | What happens |
|---|---|
| Training | Creating a model from data |
| Versioning | Tracking & comparing multiple model versions |
| Inference | Using a trained model to make predictions |
π Evaluation Metrics (Classification)
These are evaluation metrics, not algorithms. They measure how good a classification model is, derived from the confusion matrix (TP, FP, FN, TN).
π₯ Loss Function (Training)
- Measures how wrong the model is
- Used during training to improve the model
- Lower loss = better model
β Loss β Accuracy / Precision / Recall / F1
Accuracy
- How often the model is correct overall
- "Out of all predictions, how many did I get right?"
Precision
- How correct your positive predictions are
- "When the model says YES, how often is it right?"
- Formula: TP / (TP + FP)
Recall (Sensitivity)
- How many actual positives the model finds
- "Out of all real YES cases, how many did we catch?"
- Formula: TP / (TP + FN)
F1 Score
- Balance between precision and recall
- Combines both into one score
- Formula: 2 Γ (Precision Γ Recall) / (Precision + Recall)
| Focus | You care about | Exam wording |
|---|---|---|
| Precision | Fewer false positives | "Avoid false alarms" |
| Recall | Fewer false negatives | "Don't miss cases" |
| F1 | Balance of both | "Overall classification quality" |
| Accuracy | Overall correctness | "Balanced dataset, generic" |
β Regression uses MAE, MSE, RMSE.
β Clustering uses silhouette, cohesion.
| Concept | Used for | Question it answers |
|---|---|---|
| Loss | Training | How wrong am I? |
| Accuracy | Evaluation | How often am I right? |
| Precision | Evaluation | Can I trust positives? |
| Recall | Evaluation | Did I miss any positives? |
| F1 Score | Evaluation | Balance precision + recall |
- Loss β training β "how wrong"
- Accuracy β "overall correct"
- Precision β "don't cry wolf"
- Recall β "don't miss cases"
- F1 β "balance tradeoff"
| π§ͺ Real Example: Disease Detection | |
|---|---|
| Precision | Don't say disease if healthy |
| Recall | Don't miss a sick patient |
| Item | What it is |
|---|---|
| Accuracy / Precision / Recall / F1 | Evaluation metrics |
| Loss (MSE, Cross Entropy) | Training metric |
| Logistic Regression / Decision Tree / Random Forest / Neural Network | Algorithms |
π¨ Metrics do NOT train models. Metrics do NOT choose algorithms.
π Evaluation Metrics (Regression)
These metrics evaluate regression models where the output is a number (price, amount, forecast). They are NOT used for classification.
MAE β Mean Absolute Error
- Average absolute difference between predicted and actual
- Same unit as target value
- Treats all errors equally
MSE β Mean Squared Error
- Average of squared differences
- Large mistakes punished much more
- Unit is squared (less intuitive)
RMSE β Root Mean Squared Error
- Square root of MSE
- Penalizes large errors (like MSE)
- Same unit as target value (more interpretable)
| Metric | Penalizes large errors | Same unit as output | Sensitivity to outliers |
|---|---|---|---|
| MAE | β No | β Yes | Low |
| MSE | β Yes (strong) | β No (squared) | High |
| RMSE | β Yes | β Yes | High |
β Classification uses Accuracy, Precision, Recall, F1.
β Clustering uses silhouette, cohesion.
- "Predict a number" β Regression
- "Average error" β MAE
- "Penalize large errors" β MSE
- "Same units + penalize large errors" β RMSE
π Evaluation Metrics (Clustering)
Clustering is unsupervised β there are no labels. Traditional metrics like accuracy or MAE do not apply.
Silhouette Score
- How well each point fits within its cluster (cohesion)
- How well it is separated from other clusters
Cohesion (Intraβcluster similarity)
- How close together points are inside the same cluster
- High cohesion β good cluster
- Low cohesion β scattered cluster
- "Group similar items" / "No labels" β Clustering
- "Evaluate cluster quality" β Silhouette / Cohesion
- NOT classification or regression metrics
π RFM (Recency, Frequency, Monetary)
Customer Segmentation Technique
- Recency β How recently they interacted
- Frequency β How often they interact
- Monetary β How much they spend
Is the question about grouping customers by behavior, not prediction?
β RFM β Machine Learning service.
β RFM β Azure AI Vision or Language.
βοΈ Responsible AI Principles (EXAM FAVORITE)
Responsible AI principles often invalidate otherwise "technically correct" options.
Fairness
Reliability & Safety
Privacy & Security
Transparency
Accountability
Inclusiveness
Bias issue β Fairness. These are NOT the same.
βοΈ AI Solution Design Considerations
Beyond ethics β the exam also tests general AI design judgment.
Frequently Tested Considerations
- Accuracy vs Explainability
- Automation vs Humanβinβtheβloop
- Latency & response time
- Cost of inference
- Risk of misuse or overβautomation
Is full automation appropriate β or should a human remain responsible?
Generative AI Workloads
- Microsoft uses the term "Generative AI workloads"
- Includes: text generation, summarization, chat, code generation
β Azure AI Vision is NOT a GenAI workload
β AutoML is NOT a GenAI workload
β If it creates new content, it's a Generative AI workload
β Final Mental Checklist (Use This in the Exam)
- What is the input? (text, image, audio)
- What is the output? (label, number, text, location)
- Is it extracting, comparing, or generating?
- Is there structure vs raw content? (Document Intelligence vs OCR)
- Is ethics, bias, safety, or accessibility mentioned?
If you answer all 5, the correct option is usually obvious.