Search⌘ K
AI Features

Introduction to AWS Core Services for MLA-C01

Understand how foundational AWS services like EC2, S3, and IAM underpin machine learning workflows on AWS. Explore the three engineering pillars: scalability, cost-effectiveness, and security, and see how they guide service selection and configuration across the ML lifecycle stages such as data preparation, model training, deployment, and monitoring. This lesson prepares you to architect scalable, secure, and cost-efficient ML systems aligned with the AWS MLA-C01 certification requirements.

The MLA-C01 exam evaluates whether you can function as a machine learning engineer. This distinction matters because the exam goes well beyond algorithm selection. It tests your ability to architect end-to-end ML systems on AWS that are scalable, cost-effective, and secure. Every SageMaker training job you launch runs on EC2 instances. Every dataset you feed into a model lives in S3. Every API call your inference endpoint serves is governed by IAM policies. If you do not understand these foundational services, you cannot pass the exam.

This lesson maps core AWS service categories to their roles in the ML life cycle and introduces three engineering pillars that the exam uses as a lens for nearly every question. Think of this as the blueprint for the chapter ahead, which covers compute foundations, storage foundations, and serverless architectures in dedicated lessons.

Note: Amazon SageMaker is the exam's central ML service, but it delegates almost every infrastructure task to foundational AWS services like EC2, S3, IAM, and KMS. Understanding these dependencies is nonnegotiable.

The following diagram illustrates how foundational AWS services connect across a typical ML workflow, from data ingestion through deployment.

AWS ML workflow architecture with data ingestion, training and deployment
AWS ML workflow architecture with data ingestion, training and deployment

With this architecture in mind, let’s examine the three engineering pillars.

The three engineering pillars

The MLA-C01 exam organizes infrastructure decisions around three pillars. Each one maps directly to how AWS services are selected and configured within ML workflows.

Scalability

Machine learning workloads are often bursty, with periods of low activity interrupted by sudden spikes in demand. Training jobs may require large GPU clusters for a short duration, while inference endpoints can experience sharp traffic increases during events such as product launches. Amazon SageMaker endpoints support Auto ScalingA mechanism that automatically adjusts the number of active compute instances based on real-time demand metrics like CPU utilization or request count., allowing instances to be added or removed based on request count, latency, or CPU utilization. This helps ensure that capacity matches demand without overprovisioning.

The exam focuses on your ability to select appropriate scaling metrics and policies to balance cost and performance.

Cost-effectiveness

AWS provides multiple pricing levers, and the exam frequently presents scenarios in which you must choose the cheapest viable option. Three cost-optimization mechanisms appear repeatedly in exam questions.

  • S3 Intelligent-Tiering: This storage class automatically moves objects between frequent and infrequent access tiers based on usage patterns, reducing storage costs for training datasets that are accessed heavily during development but rarely afterward.

  • Managed Spot Training: SageMaker can run training jobs on Spot InstancesUnused EC2 capacity that AWS sells at up to a 90% discount compared to On-Demand pricing, with the trade-off that AWS can reclaim the instance with a two-minute warning. for fault-tolerant training jobs that support checkpointing. This can dramatically reduce costs.

  • AWS Inferentia chips: These custom-designed accelerators optimize deep learning inference at a lower cost per prediction than general-purpose GPU instances.

Security

AWS operates under a shared responsibility modelAWS secures the underlying cloud infrastructure (hardware, networking, facilities), while the customer is responsible for securing what they build in the cloud (data, IAM policies, encryption settings).. For ML engineers, this means configuring IAM policies that restrict which SageMaker notebooks can access which S3 buckets, enabling KMS encryption for data at rest, and using VPC endpointsPrivate network connections that allow AWS services to communicate without sending traffic over the public internet. to keep SageMaker traffic off the public internet.

Attention: A common exam distractor presents a scenario in which data is encrypted in S3 but transmitted unencrypted to a SageMaker training job. The correct answer typically involves enabling inter-container encryption or using VPC endpoints.

The table below summarizes how each pillar maps to specific AWS services and exam scenarios.

AWS Well-Architected Pillars for Machine Learning

Pillar

Key AWS Services

Exam-Relevant Example

Scalability

EC2 Auto Scaling, SageMaker Endpoints, Lambda

Auto-scaling a real-time inference endpoint during a traffic spike

Cost-effectiveness

S3 storage classes, Spot Instances, AWS Inferentia

Using Managed Spot Training to reduce training costs by up to 90%

Security

IAM, KMS, VPC, AWS PrivateLink

Attaching an IAM policy to restrict SageMaker notebook access to a specific S3 bucket with KMS-encrypted objects

These pillars are not isolated concerns. A well-architected ML system balances all three simultaneously, and the exam rewards candidates who can make trade-offs across them.

Mapping AWS services to ML stages

Every ML project follows a life cycle, and each stage relies on a distinct set of AWS services. Confusing which service belongs to which stage is one of the most common reasons candidates lose points on the exam.

Data preparation

Raw data lands in Amazon S3, which serves as the central data lake. From there, an AWS Glue crawler scans the data and populates the AWS Glue Data Catalog with schema metadata. This catalog enables Amazon Athena to run ad hoc SQL queries directly against S3 data without provisioning any database infrastructure. The output of these queries, such as filtered or aggregated feature sets, is written back to S3 for downstream consumption.

This stage is purely about data engineering. No model training happens here.

Model training

SageMaker training jobs provision EC2 instances behind the scenes. For deep learning workloads, GPU-backed instances like P3 (NVIDIA V100) or P4 (NVIDIA A100) are selected. Training data is pulled from S3 into the instance, and high-throughput EBS volumes provide fast local disk access during training. When the job completes, the trained model artifact is written back to S3.

The exam tests whether you can choose the right instance family. A common distractor offers a compute-optimized C5 instance for a computer vision training job when a GPU instance like P3 is the correct choice.

Model deployment

SageMaker offers three inference modes, and the exam expects you to distinguish among them.

  • Real-time endpoints: These maintain always-on instances behind a load balancer, providing low-latency responses suitable for interactive applications.

  • Serverless endpoints: These scale to zero when idle and spin up on demand, making them cost-effective for infrequent or unpredictable traffic patterns.

  • Asynchronous endpoints: These queue incoming requests and process them in the background, which is suited to large payloads or long-running inference tasks.

AWS Lambda can handle lightweight preprocessing before requests reach a SageMaker endpoint, and API Gateway exposes the endpoint as a managed REST API.

Monitoring and operations

After deployment, SageMaker Model Monitor continuously evaluates incoming data against a baseline to detect data drift or model quality degradation. CloudWatch collects logs and metrics from every component, and CloudWatch Alarms can trigger automated responses when thresholds are breached. SageMaker Pipelines orchestrates the entire workflow as a CI/CD pipeline, enabling automated retraining when model performance drops.

Practical tip: When an exam question mentions “detecting changes in input data distribution after deployment,” the answer is almost always SageMaker Model Monitor, not CloudWatch. CloudWatch handles infrastructure metrics; Model Monitor handles ML-specific metrics like feature drift.

The following mind map organizes these services by life cycle stage for quick reference.

Quick-reference taxonomy of AWS services mapped to each stage of the ML lifecycle

With the life cycle mapped, there is one more critical decision framework.

Managed AI vs. custom SageMaker models

Not every ML problem requires building a model from scratch. AWS offers fully managed AI services that handle specific use cases out of the box. Amazon Rekognition performs image and video analysis. Amazon Comprehend handles natural language processing tasks like sentiment analysis and entity extraction. Amazon Forecast generates time series predictions. These services require no ML expertise, no instance selection, and no model tuning.

The exam tests a simple decision heuristic: If a managed AI service covers 80% or more of the use case, prefer it. Custom SageMaker models that use built-in algorithms like XGBoost or custom training containers are appropriate when the problem is domain-specific, requires custom feature engineering, or demands fine-grained control over hyperparameters and training infrastructure.

The trade-off is clear. Managed AI services trade flexibility for speed and simplicity. Custom SageMaker models trade speed for full control. The exam penalizes candidates who default to building custom models when a managed service would suffice.

Note: A question that describes a company with “no ML team” and a standard use case like image classification or sentiment analysis is almost always pointing toward a managed AI service, not SageMaker.

Conclusion

The MLA-C01 exam evaluates you as a system architect who happens to specialize in machine learning. Every service choice you make on the exam should be filtered through the three engineering pillars of scalability, cost-effectiveness, and security. Foundational services like S3, EC2, IAM, and KMS are not background details; they are the infrastructure that SageMaker depends on for every training job, every endpoint, and every data access pattern. In the next lesson, we will dive into compute foundations for ML, exploring EC2 instance families, GPU selection criteria, and how Managed Spot Training works under the hood.