Copyright (c) 2026 MindMesh Academy. All rights reserved. This content is proprietary and may not be reproduced or distributed without permission.

1.3.2. SageMaker Components and When to Use Them

💡 First Principle: Each SageMaker component solves one specific problem in the ML lifecycle. The exam tests your ability to match the problem description to the correct component—not your knowledge of SageMaker's marketing materials.

ComponentLifecycle StagePrimary UseKey Exam Signals
Data WranglerData PrepVisual data prep, feature engineering"Minimal code," "visual interface," "data exploration"
Feature StoreData PrepCentralized feature repository"Reuse features," "online/offline store," "feature consistency"
Ground TruthData PrepData labeling workflows"Labeling," "annotation," "human reviewers"
Training JobsModel DevScalable model training"Train at scale," "distributed training," "GPU instances"
Automatic Model Tuning (AMT)Model DevHyperparameter optimization"Hyperparameter tuning," "Bayesian optimization," "best parameters"
JumpStartModel DevPre-trained models and solutions"Pre-trained," "transfer learning," "quick start"
ExperimentsModel DevTrack and compare training runs"Compare experiments," "track metrics," "reproducibility"
Model RegistryModel Dev → DeployModel versioning and approval"Model version," "approval workflow," "model governance"
ClarifyModel Dev + MonitorBias detection and explainability"Bias," "explainability," "SHAP values," "fairness"
DebuggerModel DevDebug training convergence issues"Vanishing gradients," "exploding gradients," "overfitting during training"
EndpointsDeploymentReal-time inference serving"Real-time predictions," "low latency," "endpoint"
Batch TransformDeploymentLarge-scale offline inference"Batch predictions," "no persistent endpoint," "offline scoring"
PipelinesDeploymentML workflow orchestration"Automate workflow," "DAG," "orchestrate steps"
NeoDeploymentModel optimization for edge"Edge deployment," "optimize model," "IoT"
Model MonitorMonitoringData/model drift detection"Drift," "data quality," "production monitoring," "baseline"
Inference RecommenderMonitoringEndpoint sizing recommendations"Rightsizing," "instance selection," "load testing"

⚠️ Exam Trap: SageMaker Clarify appears in two lifecycle stages: pre-training bias detection (Stage 2) and post-deployment monitoring (Stage 4). Similarly, Feature Store has an online store (low-latency inference lookups) and an offline store (batch training data). Questions test whether you understand these dual roles.

Reflection Question: A team notices their model's predictions disproportionately affect one demographic group. Which SageMaker component should they use, and does it matter whether the model is in development or already deployed?

Alvin Varughese
Written byAlvin Varughese
Founder15 professional certifications