What You'll Be Tested On
- Selecting model types and algorithms for classification, regression, clustering, forecasting, and anomaly detection
- Configuring SageMaker training jobs, built-in algorithms, custom containers, and distributed training
- Evaluating models with appropriate metrics and diagnosing overfitting or underfitting
- Running hyperparameter tuning jobs and selecting objective metrics
- Registering and versioning model candidates for later deployment
Key AWS Services in This Domain
Exam Tips for Domain 2
Match the metric to the business problem: precision/recall for imbalanced classification, RMSE for regression, and AUC for ranking quality.
If training is expensive and interruptible, managed spot training is often the cost-aware answer.
Use hyperparameter tuning when the model approach is chosen but performance needs systematic improvement.
Model versioning and approval belong in the model registry, not in ad hoc file names.
Practice Domain 2 Questions
Test your knowledge of ML Model Development with practice questions from our MLA-C01 question bank.
Start Practice Quiz →