The recurring wrong-answer patterns on the AIP-C01 exam, the AWS reasoning framework to apply to every question, and a decision matrix that maps common scenarios to the right service. Read this twice before exam day.
This exam introduces two new question types that SAP-C02 doesn't have. You must get the entire answer correct — no partial credit.
Given 3-5 steps, arrange them in the correct sequence. Example: "Put these RAG pipeline steps in order: embed documents, chunk documents, store in vector database, retrieve relevant chunks, generate response."
You must understand entire workflows end-to-end, not just individual steps.
Given 3-7 pairs to match. Example: "Match each AWS service to its purpose in a GenAI architecture" or "Match each symptom to its root cause."
Use process of elimination. If you're sure of 2 matches, the third may be forced by what's left.
Apply this to every question, in this order:
When you see a phrase like one of these, the right service jumps out:
| "I need to..." | AWS service(s) |
|---|---|
| Add AI to my existing app | Bedrock API + API Gateway + Lambda |
| Answer questions from company documents | Bedrock Knowledge Bases (managed RAG) |
| Build an AI that can take actions | Bedrock Agents or Strands Agents |
| Filter harmful content | Bedrock Guardrails |
| Manage prompts across teams | Bedrock Prompt Management |
| Chain multiple AI steps (no-code) | Bedrock Prompt Flows |
| Orchestrate complex logic with non-Bedrock services | AWS Step Functions |
| Deploy a custom or fine-tuned model | SageMaker AI endpoints |
| Enterprise search over internal data | Amazon Q Business or Kendra |
| Detect PII before sending to FM | Comprehend PII detection + Guardrails |
| Track model performance over time | CloudWatch + Bedrock Model Invocation Logs |
| Ensure model availability across regions | Bedrock Cross-Region Inference |
| Reduce FM costs | Model cascading + caching + prompt compression + right-sized models |
| Build a chat interface | API Gateway WebSocket + InvokeModelWithResponseStream |
| Persistent-connection MCP tool | ECS (not Lambda) |
| Approval before high-value action | Step Functions callback pattern (human-in-the-loop) |
| Document processing at scale | Step Functions + Textract + Comprehend + Bedrock |
| Bias/fairness evaluation | SageMaker Clarify or LLM-as-a-Judge |
| Explain how an agent reasoned | Bedrock Agent Tracing |
| Keep Bedrock off public internet | VPC Endpoints (PrivateLink) |
These words shift the correct answer. Highlight them mentally when you see them.
Both modify the model. Fine-tuning updates all parameters (expensive). LoRA updates a small subset (cheap, fast). LoRA is a type of parameter-efficient fine-tuning.
Guardrails = AWS-managed content filtering (a product). Safety classifiers = models trained to detect harmful content (a technique). Guardrails may use safety classifiers under the hood.
Both orchestrate multi-step workflows. Prompt Flows = no-code, Bedrock-native. Step Functions = code-based, any AWS service. Use Prompt Flows for simple prompt chains; Step Functions for complex logic with non-Bedrock services.
Bedrock Agents = fully managed, AWS-native. Strands Agents = open-source, more customizable, run on your compute. Exam prefers managed → Bedrock Agents unless customization demanded.
Semantic = vector similarity only. Hybrid = vector + keyword (BM25) combined. Hybrid usually performs better on real data, especially for exact-term queries.
On-demand = pay per token, no commitment, subject to throttling. Provisioned = reserved capacity, consistent performance, minimum commitment. Use provisioned for predictable high-volume workloads.
Overflow = input exceeds maximum, request fails. Truncation = system silently cuts off content to fit. Both lose information; different failure modes.
Model card = docs about the model itself (purpose, limits, perf). Data lineage = tracking where data came from and how it was transformed. Both serve governance at different levels.
Agents = the agent service itself (action groups, KBs, reasoning). AgentCore = infrastructure layer for deploying and scaling agents at production scale.
Q Business = enterprise assistant over internal data (S3, SharePoint, Salesforce). Q Developer = coding assistant (code gen, refactor, debug).