Skip to Content

Amazon SAA-C03: What is the most cost-effective AWS solution for short-lived batch jobs that require data storage?

Learn the optimal combination of Amazon EC2 instance types and S3 storage classes to minimize costs for an application that runs short-lived batch jobs and requires 30-day access and 2-year retention of generated data.

Table of Contents

Question

A company is migrating its data processing application to the AWS Cloud. The application processes several short-lived batch jobs that cannot be disrupted. Data is generated after each batch job is completed. The data is accessed for 30 days and retained for 2 years.

The company wants to keep the cost of running the application in the AWS Cloud as low as possible.

Which solution will meet these requirements?

A. Migrate the data processing application to Amazon EC2 Spot Instances. Store the data in Amazon S3 Standard. Move the data to Amazon S3 Glacier Instant. Retrieval after 30 days. Set an expiration to delete the data after 2 years.
B. Migrate the data processing application to Amazon EC2 On-Demand Instances. Store the data in Amazon S3 Glacier Instant Retrieval. Move the data to S3 Glacier Deep Archive after 30 days. Set an expiration to delete the data after 2 years.
C. Deploy Amazon EC2 Spot Instances to run the batch jobs. Store the data in Amazon S3 Standard. Move the data to Amazon S3 Glacier Flexible Retrieval after 30 days. Set an expiration to delete the data after 2 years.
D. Deploy Amazon EC2 On-Demand Instances to run the batch jobs. Store the data in Amazon S3 Standard. Move the data to Amazon S3 Glacier Deep Archive after 30 days. Set an expiration to delete the data after 2 years.

Answer

C. Deploy Amazon EC2 Spot Instances to run the batch jobs. Store the data in Amazon S3 Standard. Move the data to Amazon S3 Glacier Flexible Retrieval after 30 days. Set an expiration to delete the data after 2 years.

Explanation

The most cost-effective solution that meets all the requirements is Option C:

  • Deploy Amazon EC2 Spot Instances to run the batch jobs.
  • Store the data in Amazon S3 Standard.
  • Move the data to Amazon S3 Glacier Flexible Retrieval after 30 days.
  • Set an expiration to delete the data after 2 years.

EC2 Spot Instances provide the lowest cost compute option that is suitable for short-lived batch jobs that can tolerate interruptions. The application processes several batch jobs that cannot be disrupted, making Spot Instances an appropriate choice. On-Demand Instances are not needed since a small interruption is acceptable.

For the data storage, S3 Standard provides low latency and high throughput, which is ideal for frequently accessed data in the first 30 days. After 30 days, S3 Glacier Flexible Retrieval offers lower costs for data that is accessed less frequently but still required within minutes when needed over the 2 year retention period.

S3 Glacier Deep Archive is even lower cost, but has a longer retrieval time of 12-48 hours, making Flexible Retrieval a better choice for the 2 year retention where faster access may still be needed on occasion. Setting a lifecycle policy to delete the data after 2 years avoids unnecessary ongoing storage costs.

So in summary, using low-cost Spot Instances for compute, S3 Standard for the initial 30 days of storage, S3 Glacier Flexible Retrieval for the remainder of the 2 year retention, and automatically deleting data after 2 years provides the most cost-effective solution while meeting all stated requirements.

Amazon AWS Certified Solutions Architect – Associate SAA-C03 certification exam assessment practice question and answer (Q&A) dump including multiple choice questions (MCQ) and objective type questions, with detail explanation and reference available free, helpful to pass the Amazon AWS Certified Solutions Architect – Associate SAA-C03 exam and earn Amazon AWS Certified Solutions Architect – Associate SAA-C03 certification.