Domain 5: Evaluation and Troubleshooting - Exam Cues

”Least Operational Overhead” Winners

This specific list of services appears repeatedly as the “correct” answer when the question asks for “Least Operational Overhead” or “Least Development Effort.”

CategoryThe “Winner” ServiceWhy?
OrchestrationStep FunctionsManaged state machine; better than managing Lambda code/retries.
Text ExtractionTextract / BDAManaged OCR; better than custom libraries.
Search/RAGKnowledge BasesFully managed RAG pipeline; better than building your own.
Apps/ChatAmazon Q BusinessOut-of-the-box chat app; better than building a React UI.
EvaluationLLM-as-a-judgeAutomated scoring; better than manual human review.
Data QualityGlue Data QualityManaged rules; better than writing Python validation scripts.
Priv. NetworkVPC EndpointsManaged PrivateLink; better than VPN/DirectConnect for service access.
ClassificationComprehendManaged classifier; better than fine-tuning a model for simple tags.

Exam Cues

If you see…Think…
”Compare summaries” + “Automated”ROUGE / BLEU (via Bedrock Evaluation)
“Validate new model” + “No user impact”Shadow Testing (SageMaker)
“Drift” or “Degraded accuracy”Model Monitor
”Audit all prompts/responses”Invocation Logging (to S3)
“Latency alert”CloudWatch Alarm (on InvocationLatency)
“Subjective quality” + “Scale”LLM-as-a-judge (Automated Evaluation)
“Ground Truth” + “High Risk”Human Evaluation (Own Team)