SageMaker Training JobsSageMaker Pipelines
#1Both train models, so candidates reach for Training Jobs without checking whether the scenario involves a repeatable multi-step workflow.
Deciding signal
A SageMaker Training Job is a single managed compute run that takes data from S3, runs a training algorithm on ML instances, and writes model artifacts back to S3. It answers one question: produce a trained model artifact. SageMaker Pipelines is a CI/CD orchestration service for ML workflows — it chains preprocessing, training, evaluation, approval gates, and registration into a repeatable, version-controlled DAG. Pipelines is the right answer when the scenario describes automating the end-to-end ML lifecycle, running the same workflow on new data batches, or enforcing quality gates before model registration.
Quick check
Is this a one-time compute run to produce a model artifact (Training Job), or an automated repeatable workflow with multiple steps including evaluation and approval (Pipelines)?
Why it looks right
Training Jobs are the fundamental SageMaker compute primitive and candidates apply them to pipeline orchestration questions. SageMaker Pipelines is the specific answer when the workflow has multiple stages that must be automated.
SageMaker Real-Time EndpointsSageMaker Batch TransformSageMaker Async Inference
#2All three run inference using a deployed model, so candidates default to endpoints for all prediction scenarios.
Deciding signal
Real-Time Endpoints serve synchronous inference requests — the caller waits for the response, latency matters, and the endpoint must stay running. Best for user-facing applications. Batch Transform runs inference on large datasets stored in S3 without a persistent endpoint — it spins up instances, scores all records, and shuts down. Best for offline scoring jobs where latency is irrelevant and cost efficiency matters. Async Inference accepts requests with large payloads or long processing times and returns results via S3 notification — it queues requests and scales the endpoint down to zero when idle. Best for large document or video processing where synchronous timeout limits would be exceeded.
Quick check
Is this a user-facing real-time request requiring low latency (Real-Time Endpoint), scoring a large S3 dataset offline (Batch Transform), or processing large payloads asynchronously with no persistent endpoint (Async Inference)?
Why it looks right
Real-Time Endpoints are the default inference answer. Batch Transform is overlooked when the scenario clearly describes an offline dataset — candidates try to use endpoints when Batch Transform is the appropriate lower-cost option.
SageMaker Feature StoreSageMaker Data WranglerAWS Glue
#3All three process or store data for ML, so candidates conflate the data preparation and feature management steps.
Deciding signal
Data Wrangler is a visual data preparation tool inside SageMaker Studio — it imports data from S3, Redshift, or other sources, applies transformations with a point-and-click UI, and exports the prepared dataset. It is a data preparation tool, not a feature store. Feature Store is a repository for ML features — engineered values that are stored with consistent naming, versioned, and reused across training and inference. It has an online store (low-latency lookup) and offline store (S3-backed for training). Glue is a serverless ETL service for transforming and loading data into S3, databases, or data warehouses — it is more general-purpose and not ML-specific. When the scenario involves preparing a dataset for a specific training run, Data Wrangler. When it involves storing and reusing computed features across multiple models, Feature Store.
Quick check
Is this preparing a dataset for a specific training run (Data Wrangler), storing and reusing features across models and training/inference (Feature Store), or general data transformation to S3 or a warehouse (Glue)?
Why it looks right
Glue is the familiar data processing answer and candidates apply it to SageMaker data preparation scenarios. Data Wrangler and Feature Store are SageMaker-native tools with specific roles that Glue does not fill.
SageMaker Ground TruthAmazon A2I (Augmented AI)
#4Both involve human review of ML outputs, so candidates treat them as alternatives for any "human-in-the-loop" question.
Deciding signal
Ground Truth manages the labeling of training data — images, text, video — using a workforce (Amazon Mechanical Turk, private team, or third party). The output is a labeled dataset for training a model. A2I (Augmented AI) inserts human review into a production ML prediction workflow when a model returns low-confidence predictions. It integrates with Textract, Rekognition, and custom models. The output is a reviewed prediction with a human override where needed. The timing is the distinction: Ground Truth is pre-training (producing training data); A2I is post-deployment (reviewing production predictions).
Quick check
Is the goal to produce labeled training data before model training (Ground Truth), or to review low-confidence model predictions in a production inference pipeline (A2I)?
Why it looks right
Both involve a human workforce reviewing ML-related content. Candidates conflate "humans reviewing model outputs" regardless of whether it is a labeling task or a production review task.
Amazon EMRAWS GlueAmazon Athena
#5All three process large datasets and appear in ML data pipeline questions.
Deciding signal
EMR provisions a managed cluster of EC2 instances running Hadoop, Spark, Hive, Presto, or other distributed frameworks. You control the cluster size, instance types, and framework configuration — it is the right answer when you need full control of the Spark environment, custom libraries, or processing that exceeds Glue limits. Glue is serverless ETL: you define jobs in PySpark or Python and Glue manages the compute — no cluster management required. Athena is a serverless SQL query engine that reads data directly from S3 — there is no data movement or ETL step. The distinction is the operation: ad-hoc SQL queries on S3 data (Athena), managed ETL jobs without cluster management (Glue), or custom Spark processing with full cluster control (EMR).
Quick check
Is this SQL queries directly on S3 with no ETL (Athena), serverless managed ETL transformations (Glue), or custom Spark processing on a managed cluster you control (EMR)?
Why it looks right
EMR and Glue both run Spark and candidates conflate them. Glue is the correct answer when the exam description emphasizes "no infrastructure management" for ETL. EMR is correct when custom cluster configuration or non-Glue Spark functionality is required.
SageMaker CanvasSageMaker Autopilot
#6Both automate model training, so candidates apply Canvas as the "easier" option without checking who the intended user is.
Deciding signal
Canvas is a visual, no-code tool for business analysts. It imports data, automatically trains a model, and produces predictions — no ML expertise required. It is designed for non-technical users. Autopilot is AutoML for ML practitioners: it trains and evaluates many candidate model/algorithm combinations, produces leaderboard results, and generates explainability reports (SageMaker Clarify integration). Autopilot exposes the full training configuration and notebooks so practitioners can inspect and reproduce what was done. The user is the key: business analysts using a point-and-click UI (Canvas) versus ML practitioners who want automated training with visibility into the process (Autopilot).
Quick check
Are the users non-technical business analysts who need no ML expertise (Canvas), or ML practitioners who want automated model selection with explainability and transparency (Autopilot)?
Why it looks right
Canvas sounds like the simpler option and candidates apply it when automated training is mentioned without checking whether transparency and explainability are required — which points to Autopilot.
SageMaker Built-in AlgorithmsBring Your Own Container (BYOC)SageMaker JumpStart
#7All three are ways to run training or inference in SageMaker, so candidates apply built-in algorithms universally.
Deciding signal
Built-in algorithms (XGBoost, Linear Learner, DeepAR, BlazingText, etc.) are optimized AWS-managed training algorithm containers — you provide training data and hyperparameters, and SageMaker handles the rest. They are the right answer for standard ML tasks where an AWS-optimized implementation of a known algorithm is available. BYOC lets you package a custom training or inference script in a Docker container and run it on SageMaker infrastructure — required when you need a custom framework, library, or algorithm not available as a built-in. JumpStart provides a hub of pre-trained foundation models and ML models that can be deployed or fine-tuned — you are not training from scratch. JumpStart is the right answer when a pre-trained model is sufficient and you do not want to start training from scratch.
Quick check
Is a standard AWS-optimized algorithm sufficient (Built-in), does the workload need a custom framework packaged in Docker (BYOC), or is a pre-trained model being deployed or fine-tuned (JumpStart)?
Why it looks right
Built-in algorithms are the default SageMaker training answer. BYOC is the correct answer when the scenario explicitly requires a library or framework not available in AWS built-ins.
SageMaker Hyperparameter Tuning (HPO)SageMaker Experiments
#8Both involve running multiple training jobs and comparing results, so candidates conflate the mechanism with the analysis layer.
Deciding signal
SageMaker Hyperparameter Tuning automatically searches a defined hyperparameter space — learning rate, batch size, regularization — across multiple training jobs to find the combination that maximizes your objective metric. It manages the search strategy (Bayesian, random, grid) and selects the best model. SageMaker Experiments provides a framework for organizing, tracking, and comparing training runs — you log metrics, parameters, and artifacts from any training job and visualize them in a comparison view. HPO automates the search; Experiments provides observability and comparison across any training runs, including manual ones. When the scenario asks how to automatically find the best hyperparameters, HPO. When it asks how to track and compare runs across an ML project, Experiments.
Quick check
Is the goal to automatically search for optimal hyperparameters (HPO), or to log, organize, and compare metrics across multiple training runs (Experiments)?
Why it looks right
HPO is the more prominent SageMaker feature. SageMaker Experiments is correct when the scenario describes tracking and comparing training runs — a different problem from automated hyperparameter search.
SageMaker ClarifySageMaker Model Monitor
#9Both analyze model behavior, so candidates apply Model Monitor to all model quality questions.
Deciding signal
Clarify analyzes models for bias in training data or predictions and provides feature attribution (SHAP values) to explain individual predictions. It runs as a processing job during training and evaluation to assess fairness and explainability. Model Monitor continuously monitors deployed models in production for data quality drift (input feature distribution changes), model quality drift (accuracy degradation), bias drift, and feature attribution drift. It generates alerts when metrics exceed thresholds. Clarify is used at training time to understand the model; Model Monitor is used in production to detect when the model or data has changed. Both can work together — Clarify can baseline the bias metrics that Model Monitor then tracks.
Quick check
Is this analyzing bias and feature importance during model development (Clarify), or monitoring a deployed model in production for data or model quality drift (Model Monitor)?
Why it looks right
Model Monitor is the operational monitoring tool and candidates apply it to explainability questions. Clarify is specifically for understanding bias and feature attribution — a distinct analytical step from production drift monitoring.
Amazon Kinesis Data StreamsAmazon Kinesis Data Firehose
#10Both ingest streaming data for ML pipelines, so candidates treat them interchangeably.
Deciding signal
Kinesis Data Streams requires you to build consumer applications that read records in real time. It supports multiple consumers, configurable retention, and replay — giving full control over how streaming data is processed before it reaches a model or feature store. Kinesis Data Firehose automatically delivers data to a destination (S3, Redshift, OpenSearch) with optional Lambda transformation — no consumer code required. For ML pipelines, Data Streams is the right answer when the scenario involves real-time feature computation before writing to a Feature Store. Firehose is the right answer when streaming data should land in S3 for batch training with minimal code.
Quick check
Does the ML pipeline need real-time processing with custom consumer logic and replay (Data Streams), or automatic delivery of streaming data to S3 or another store without consumer management (Firehose)?
Why it looks right
Firehose is the easier option and candidates apply it to real-time streaming ML scenarios. Data Streams is correct when the scenario requires custom processing logic or replay capability before the data reaches an ML system.
10 MLS-C01 questions. Pattern-tagged with trap analysis. Free, no signup required.
Start MLS-C01 Mini-Trainer →