9.2. High-Frequency Traps and Exam Patterns
These are the distinctions the exam tests most reliably — the questions that catch candidates who learned features in isolation but did not internalize the differences.
The 10 most important distinctions for AB-730:
| Topic | The Trap | The Correct Understanding |
|---|---|---|
| Fabrication frequency | "Hallucinations are rare edge cases" | Fabrications are a structural characteristic — they can occur on any topic without grounding |
| Save vs. Schedule | "I saved the prompt so it will run automatically" | Saving ≠ automating. Schedule is the feature that runs prompts automatically |
| Chat vs. Agent | "An agent is a smarter version of Copilot Chat" | They are different experiences with different purposes. Agents are scoped; chat is general |
| Copilot and OpenAI | "My data goes to OpenAI when I use Copilot" | Data stays in your Microsoft 365 tenant. Azure OpenAI is used, not OpenAI consumer services |
| Permission inheritance | "Copilot can access files I can't see" | Copilot inherits your exact M365 permissions — it cannot surface restricted content |
| DLP and output | "DLP only restricts what Copilot can access" | DLP also restricts what Copilot outputs — it filters responses, not just access |
| Meeting recap vs. transcription | "Recap and transcription are the same thing" | Transcription = verbatim record. Recap = AI synthesis of decisions and action items |
| Copilot Pages vs. SharePoint | "Copilot Pages is just another SharePoint page" | Pages is a collaborative AI canvas. SharePoint pages are static publishing artifacts |
| Copilot memory | "Copilot automatically remembers all past conversations" | Memory only retains what you explicitly configure it to remember |
| Prompt injection | "Prompt injection only affects developers" | Any user who asks Copilot to process external content is exposed to injection risk |
Recognizing scenario question patterns:
AB-730 scenarios follow predictable patterns. Once you recognize the pattern, the correct answer becomes clearer.
Pattern 1 — "Which feature should they use?" → The question describes a workflow goal. Match it to the right Copilot feature using the decision frameworks from Phases 4–8.
Pattern 2 — "What went wrong?" → Describes a suboptimal Copilot output. Diagnose the root cause: missing GCSF elements, wrong grounding source, wrong feature chosen, or AI risk not mitigated.
Pattern 3 — "What should they do next?" → After Copilot generates output. The answer almost always involves human review before action — especially for external, high-stakes, or factual content.
Pattern 4 — "What is the risk?" → Describes a business scenario involving Copilot. Identify the correct risk category: fabrication, injection, over-reliance, sensitive data exposure, or permission boundary issue.
Pattern 5 — "What is the best prompt?" → Presents multiple prompts of varying quality. The correct answer has all four GCSF components; wrong answers are missing context, source, or format specificity.