Introduction to AWS Core Services for MLA-C01
Understand how foundational AWS services like EC2, S3, and IAM underpin machine learning workflows on AWS. Explore the three engineering pillars: scalability, cost-effectiveness, and security, and see how they guide service selection and configuration across the ML lifecycle stages such as data preparation, model training, deployment, and monitoring. This lesson prepares you to architect scalable, secure, and cost-efficient ML systems aligned with the AWS MLA-C01 certification requirements.
The MLA-C01 exam evaluates whether you can function as a machine learning engineer. This distinction matters because the exam goes well beyond algorithm selection. It tests your ability to architect end-to-end ML systems on AWS that are scalable, cost-effective, and secure. Every SageMaker training job you launch runs on EC2 instances. Every dataset you feed into a model lives in S3. Every API call your inference endpoint serves is governed by IAM policies. If you do not understand these foundational services, you cannot pass the exam.
This lesson maps core AWS service categories to their roles in the ML life cycle and introduces three engineering pillars that the exam uses as a lens for nearly every question. Think of this as the blueprint for the chapter ahead, which covers compute foundations, storage foundations, and serverless architectures in dedicated lessons.
Note: Amazon SageMaker is the exam's central ML service, but it delegates almost every infrastructure task to foundational AWS services like EC2, S3, IAM, and KMS. Understanding these dependencies is nonnegotiable.
The following diagram illustrates how foundational AWS services connect across a typical ML workflow, from data ingestion through deployment.
With this architecture in mind, let’s examine the three engineering pillars.
The three engineering pillars
The MLA-C01 exam organizes infrastructure decisions around three pillars. Each one maps directly to how AWS services are selected and configured within ML workflows.
Scalability
Machine learning workloads are often bursty, with periods of low activity interrupted by sudden spikes in demand. Training jobs may require large GPU clusters for a short duration, while inference endpoints can experience sharp traffic increases during events such as product launches. Amazon SageMaker endpoints support
The exam focuses on your ability to select appropriate scaling metrics and policies to balance cost and performance.
Cost-effectiveness
AWS provides multiple pricing levers, and the exam frequently presents scenarios in which you must choose the cheapest viable option. Three cost-optimization mechanisms appear repeatedly in exam questions.
S3 Intelligent-Tiering: This storage class automatically moves objects between frequent and infrequent access tiers based on usage patterns, reducing storage costs for training datasets that are accessed heavily during development but rarely afterward.
Managed Spot Training: SageMaker can run training jobs on
for fault-tolerant training jobs that support checkpointing. This can dramatically reduce costs.Spot Instances Unused EC2 capacity that AWS sells at up to a 90% discount compared to On-Demand pricing, with the trade-off that AWS can reclaim the instance with a two-minute warning. AWS Inferentia chips: These custom-designed accelerators optimize deep learning inference at a lower cost per prediction than general-purpose GPU instances.
Security
AWS operates under a
Attention: A common exam distractor presents a scenario in which data is encrypted in S3 but transmitted unencrypted to a SageMaker training job. The correct answer typically involves enabling inter-container encryption or using VPC endpoints.
The table below summarizes how each pillar maps to specific AWS services and exam scenarios.
AWS Well-Architected Pillars for Machine Learning
Pillar | Key AWS Services | Exam-Relevant Example |
Scalability | EC2 Auto Scaling, SageMaker Endpoints, Lambda | Auto-scaling a real-time inference endpoint during a traffic spike |
Cost-effectiveness | S3 storage classes, Spot Instances, AWS Inferentia | Using Managed Spot Training to reduce training costs by up to 90% |
Security | IAM, KMS, VPC, AWS PrivateLink | Attaching an IAM policy to restrict SageMaker notebook access to a specific S3 bucket with KMS-encrypted objects |
These pillars are not isolated concerns. A well-architected ML system balances all three simultaneously, and the exam rewards candidates who can make trade-offs across them.
Mapping AWS services to ML stages
Every ML project follows a life cycle, and each stage relies on a distinct set of AWS services. Confusing which service belongs to which stage is one of the most common reasons candidates lose points on the exam.
Data preparation
Raw data lands in Amazon S3, which serves as the central data lake. From there, an AWS Glue crawler scans the data and populates the AWS Glue Data Catalog with schema metadata. This catalog enables Amazon Athena to run ad hoc SQL queries directly against S3 data without provisioning any database infrastructure. The output of these queries, such as filtered or aggregated feature sets, is written back to S3 for downstream consumption.
This stage is purely about data engineering. No model training happens here.
Model training
SageMaker training jobs provision EC2 instances behind the scenes. For deep learning workloads, GPU-backed instances like P3 (NVIDIA V100) or P4 (NVIDIA A100) are selected. Training data is pulled from S3 into the instance, and high-throughput EBS volumes provide fast local disk access during training. When the job completes, the trained model artifact is written back to S3.
The exam tests whether you can choose the right instance family. A common distractor offers a compute-optimized C5 instance for a computer vision training job when a GPU instance like P3 is the correct choice.
Model deployment
SageMaker offers three inference modes, and the exam expects you to distinguish among them.
Real-time endpoints: These maintain always-on instances behind a load balancer, providing low-latency responses suitable for interactive applications.
Serverless endpoints: These scale to zero when idle and spin up on demand, making them cost-effective for infrequent or unpredictable traffic patterns.
Asynchronous endpoints: These queue incoming requests and process them in the background, which is suited to large payloads or long-running inference tasks.
AWS Lambda can handle lightweight preprocessing before requests reach a SageMaker endpoint, and API Gateway exposes the endpoint as a managed REST API.
Monitoring and operations
After deployment, SageMaker Model Monitor continuously evaluates incoming data against a baseline to detect data drift or model quality degradation. CloudWatch collects logs and metrics from every component, and CloudWatch Alarms can trigger automated responses when thresholds are breached. SageMaker Pipelines orchestrates the entire workflow as a CI/CD pipeline, enabling automated retraining when model performance drops.
Practical tip: When an exam question mentions “detecting changes in input data distribution after deployment,” the answer is almost always SageMaker Model Monitor, not CloudWatch. CloudWatch handles infrastructure metrics; Model Monitor handles ML-specific metrics like feature drift.
The following mind map organizes these services by life cycle stage for quick reference.
With the life cycle mapped, there is one more critical decision framework.
Managed AI vs. custom SageMaker models
Not every ML problem requires building a model from scratch. AWS offers fully managed AI services that handle specific use cases out of the box. Amazon Rekognition performs image and video analysis. Amazon Comprehend handles natural language processing tasks like sentiment analysis and entity extraction. Amazon Forecast generates time series predictions. These services require no ML expertise, no instance selection, and no model tuning.
The exam tests a simple decision heuristic: If a managed AI service covers 80% or more of the use case, prefer it. Custom SageMaker models that use built-in algorithms like XGBoost or custom training containers are appropriate when the problem is domain-specific, requires custom feature engineering, or demands fine-grained control over hyperparameters and training infrastructure.
The trade-off is clear. Managed AI services trade flexibility for speed and simplicity. Custom SageMaker models trade speed for full control. The exam penalizes candidates who default to building custom models when a managed service would suffice.
Note: A question that describes a company with “no ML team” and a standard use case like image classification or sentiment analysis is almost always pointing toward a managed AI service, not SageMaker.
Conclusion
The MLA-C01 exam evaluates you as a system architect who happens to specialize in machine learning. Every service choice you make on the exam should be filtered through the three engineering pillars of scalability, cost-effectiveness, and security. Foundational services like S3, EC2, IAM, and KMS are not background details; they are the infrastructure that SageMaker depends on for every training job, every endpoint, and every data access pattern. In the next lesson, we will dive into compute foundations for ML, exploring EC2 instance families, GPU selection criteria, and how Managed Spot Training works under the hood.