The AWS Certified Machine Learning Engineer Associate (MLA-C01) is a SageMaker exam. That sounds reductive, but it's the single most useful framing for anyone starting to study. Roughly 60-80% of scored questions touch SageMaker in some form: pipelines, training jobs, feature stores, model registry, endpoint deployment. If you walk in without deep SageMaker fluency, the other AWS services won't save you.

This guide covers the exam structure, breaks down what each domain actually tests, and lays out a study approach built around the services that carry the most weight.

Exam Format

The MLA-C01 has 65 questions and a 130-minute time limit. You need a scaled score of 720 out of 1,000 to pass. Question types include standard multiple-choice (one correct answer from four), multiple-response (two or more correct from five or more options), ordering questions, and matching questions. The ordering and matching formats are newer to AWS exams and tend to appear in the deployment and orchestration domain, where you're asked to sequence pipeline steps or match services to their roles.

Fifteen of the 65 questions are unscored. AWS uses them to evaluate questions for future exams. You won't know which ones they are, so treat every question like it counts.

The exam costs $150 USD. AWS recommends at least one year of hands-on experience building ML solutions on their platform before attempting it.

The Four Domains

Domain 1: Data Preparation (28%)

This is the largest domain, and it's where many candidates underperform because they focus their study on model training instead. Data preparation on the MLA-C01 means knowing how to get data from source systems into a format that SageMaker can train on, at scale, reliably.

The specific services you need cold:

  • AWS Glue for ETL jobs, crawlers, and the Data Catalog. Know when Glue is the right choice versus running Spark on EMR. Glue handles structured and semi-structured data from S3, RDS, and Redshift well; EMR is for when you need more control over the Spark cluster.
  • SageMaker Feature Store for feature management. The exam tests whether you understand why Feature Store exists: it keeps training and inference features consistent, which prevents train-serve skew. It also enables feature reuse across teams and models.
  • SageMaker Data Wrangler for visual data preparation. Know its capabilities but also its limits; it's a UI tool for data exploration and transformation, not a replacement for Glue on large-scale ETL.
  • Data formats and storage. Parquet versus CSV versus RecordIO. When to use each. SageMaker's built-in algorithms often expect RecordIO or CSV; understanding format conversion is testable.

Feature engineering questions show up here too. Expect questions about handling missing values, encoding categorical variables, scaling numerical features, and when to apply specific transformations. The exam frames these as AWS-service questions, not statistics questions: which SageMaker processing job or Glue transform accomplishes the goal?

Domain 2: ML Model Development (26%)

Model development on this exam is less about ML theory and more about which SageMaker tools handle each stage of the training workflow. You should know:

  • Built-in algorithms. XGBoost, Linear Learner, BlazingText, k-NN, LightGBM, and others. The exam tests when to select each one. XGBoost for tabular data with complex feature interactions. Linear Learner for regression and binary classification on large datasets. BlazingText for text classification and Word2Vec. Know the input formats each algorithm expects.
  • SageMaker Training Jobs. How to configure instance types, distributed training with data parallelism versus model parallelism, and spot instance training for cost optimization. The cost angle matters; several questions test whether you can pick the approach that balances performance against spend.
  • Hyperparameter tuning. SageMaker Automatic Model Tuning runs Bayesian optimization by default. Know how to set objective metrics, define hyperparameter ranges, and interpret tuning job results.
  • Model evaluation. Accuracy, precision, recall, F1, AUC-ROC. The exam won't ask you to calculate these from scratch, but it will give you a confusion matrix or metric set and ask which model performs better for a given use case. An imbalanced fraud detection dataset where accuracy is 99% but recall is 12% is a classic setup.

SageMaker Experiments and Model Registry appear in this domain. Model Registry is how you version, catalog, and approve models before they move to production. Expect at least a few questions on the approval workflow.

Domain 3: Deployment and Orchestration (22%)

This domain tests whether you can get a trained model into production and keep the pipeline that built it running automatically. Key areas:

  • Endpoint types. Real-time endpoints for synchronous inference, batch transform for large offline prediction jobs, asynchronous endpoints for payloads that take longer to process, and serverless endpoints for intermittent traffic. The exam gives you a traffic pattern and expects you to pick the right endpoint type.
  • SageMaker Pipelines. This is the CI/CD backbone for ML on AWS. You define steps (processing, training, evaluation, registration, deployment) as a directed acyclic graph. Expect questions on step dependencies, conditional logic, and how Pipelines integrates with Model Registry.
  • Multi-model and multi-container endpoints. When you have dozens of models serving different customer segments, multi-model endpoints let you host them on shared infrastructure. Know the tradeoffs: cold start latency on less-frequently-called models versus cost savings from consolidation.
  • Infrastructure as code. CloudFormation and CDK for deploying ML infrastructure. Step Functions for orchestrating workflows that extend beyond SageMaker.

Domain 4: Monitoring, Maintenance, and Security (24%)

The second-largest domain. It catches candidates who studied training and deployment but skipped the operational side.

  • SageMaker Model Monitor. Detects data drift, model quality degradation, bias drift, and feature attribution drift. Know the four monitoring types and when to configure each. Data drift is the most commonly tested: your production input distribution shifts away from your training distribution, and model performance degrades even though nothing in the code changed.
  • CloudWatch integration. Metrics, alarms, and logs for SageMaker endpoints. Invocation count, model latency, CPU/GPU utilization. The exam tests whether you can set up alerts that trigger retraining when performance drops below a threshold.
  • Security. IAM roles for SageMaker execution, VPC configurations for training jobs and endpoints, encryption at rest and in transit with KMS, and network isolation. VPC endpoint questions are common: how to keep SageMaker traffic off the public internet.
  • Cost optimization. Spot instances for training, auto-scaling for endpoints, choosing the right instance family. The exam presents scenarios where you need to reduce cost without breaking SLA commitments.

What to Study First

Start with SageMaker. Specifically, go through the SageMaker Developer Guide sections on training jobs, built-in algorithms, and endpoints. Then move to Pipelines and Model Monitor. These four components cover roughly half the exam by themselves.

After SageMaker, pick up Glue and Feature Store for the data preparation domain. Then fill in Step Functions, CloudFormation, IAM policies, and KMS for the orchestration and security questions.

A reasonable timeline for someone with existing ML knowledge and some AWS experience:

  • Weeks 1-2: SageMaker core: training jobs, built-in algorithms, hyperparameter tuning, model evaluation. Run at least one training job end-to-end in the console or SDK.
  • Weeks 3-4: Data preparation. Glue ETL, Feature Store, data formats, processing jobs. Feature engineering patterns.
  • Weeks 5-6: Deployment and orchestration. Endpoint types, Pipelines, Model Registry, Step Functions. Set up a simple pipeline that trains, evaluates, and conditionally deploys.
  • Weeks 7-8: Monitoring and security. Model Monitor setup, CloudWatch alarms, IAM roles, VPC configuration. Take full-length practice exams and review every wrong answer.

If you already work with SageMaker daily, compress this to 4-5 weeks. If you're coming from a data science background with minimal AWS experience, budget 10 weeks and spend extra time on the infrastructure and security topics.

Common Mistakes

Three patterns show up repeatedly in post-exam reports from candidates who failed:

Underestimating Domain 1. Data preparation is 28% of the exam, but many candidates treat it as the "easy" domain and allocate the least study time. Glue ETL and Feature Store questions require specific service knowledge, not general data engineering intuition.

Studying ML theory instead of AWS implementation. You don't need to derive backpropagation. You need to know that SageMaker's XGBoost implementation expects CSV or LibSVM input, supports Pipe mode for streaming large datasets, and can run distributed training across multiple instances. The exam tests AWS service knowledge, not textbook ML.

Ignoring Model Monitor. Candidates who work primarily on model development often skip monitoring entirely. Four types of monitoring, each with specific configuration requirements, across 24% of the exam. Don't skip it.

TechPrep AWS ML Engineer

2,500 practice questions across all 4 domains. 1,500 multiple-choice covering SageMaker pipelines, Glue ETL, feature stores, endpoint deployment, and Model Monitor. 1,000 rapid-fire drills on instance types, built-in algorithms, data formats, and service-specific details.

Test Day Strategy

With 65 questions in 130 minutes, you have exactly 2 minutes per question. That sounds generous until you hit a multi-response question with six options and a detailed scenario. Flag long questions and come back to them. Unlike the ASVAB's adaptive format, AWS certification exams let you review and change answers.

Read the scenario carefully before looking at the answer choices. AWS exam questions often include details that rule out two options immediately. A question that mentions "intermittent traffic with unpredictable spikes" is steering you toward serverless or asynchronous endpoints, not real-time. A question about "retraining when data distribution changes" is pointing at Model Monitor with CloudWatch alarms, not a cron job.

When two answers both seem correct, look for the one that's more operationally sound. AWS consistently favors managed services over custom solutions, automation over manual processes, and least-privilege security over convenience. If one answer uses a native SageMaker feature and the other involves stitching together Lambda functions and custom scripts, the SageMaker answer is almost always right.

Anthony C. Perry

M.S. Computer Science, M.S. Kinesiology. USAF veteran and founder of Meridian Labs. ORCID