A company has stored 10 TB of log files in Apache Parquet format in an Amazon S3 bucket The company occasionally needs to use SQL to analyze the log files Which solution will meet these requirements MOST cost-effectively?
A. Create an Amazon Aurora MySQL database Migrate the data from the S3 bucket intoAurora by using AWS Database Migration Service (AWS DMS) Issue SQL statements tothe Aurora database. B. Create an Amazon Redshift cluster Use Redshift Spectrum to run SQL statementsdirectly on the data in the S3 bucket C. Create an AWS Glue crawler to store and retrieve table metadata from the S3 bucketUse Amazon Athena to run SQL statements directly on the data in the S3 bucket D. Create an Amazon EMR cluster Use Apache Spark SQL to run SQL statements directlyon the data in the S3 bucket
Answer: C Explanation: AWS Glue is a serverless data integration service that can crawl, catalog,and prepare data for analysis. AWS Glue can automatically discover the schema andpartitioning of the data stored in Apache Parquet format in S3, and create a table in theAWS Glue Data Catalog. Amazon Athena is a serverless interactive query service that canrun SQL queries directly on data in S3, without requiring any data loading ortransformation. Athena can use the table metadata from the AWS Glue Data Catalog toquery the data in S3. By using AWS Glue and Athena, you can analyze the log files in S3most cost-effectively, as you only pay for the resources consumed by the crawler and thequeries, and you do not need to provision or manage any servers or clusters.References:AWS GlueAmazon AthenaAnalyzing Data in S3 using Amazon Athena
Question # 352
A pharmaceutical company is developing a new drug. The volume of data that the company generates has grown exponentially over the past few months. The company'sresearchers regularly require a subset of the entire dataset to be immediately available with minimal lag. However the entire dataset does not need to be accessed on a daily basis. All the data currently resides in on-premises storage arrays, and the company wants to reduce ongoing capital expenses. Which storage solution should a solutions architect recommend to meet these requirements?
A. Run AWS DataSync as a scheduled cron job to migrate the data to an Amazon S3bucket on an ongoing basis. B. Deploy an AWS Storage Gateway file gateway with an Amazon S3 bucket as the targetstorage Migrate the data to the Storage Gateway appliance. C. Deploy an AWS Storage Gateway volume gateway with cached volumes with anAmazon S3 bucket as the target storage. Migrate the data to the Storage Gatewayappliance. D. Configure an AWS Site-to-Site VPN connection from the on-premises environment toAWS. Migrate data to an Amazon Elastic File System (Amazon EFS) file system.
Answer: C Explanation: AWS Storage Gateway is a hybrid cloud storage service that allows you to seamlessly integrate your on-premises applications with AWS cloud storage. VolumeGateway is a type of Storage Gateway that presents cloud-backed iSCSI block storagevolumes to your on-premises applications. Volume Gateway operates in either cache modeor stored mode. In cache mode, your primary data is stored in Amazon S3, while retainingyour frequently accessed data locally in the cache for low latency access. In stored mode,your primary data is stored locally and your entire dataset is available for low latencyaccess on premises while also asynchronously getting backed up to Amazon S3.For the pharmaceutical company’s use case, cache mode is the most suitable option, as itmeets the following requirements:It reduces the need to scale the on-premises storage infrastructure, as most of thedata is stored in Amazon S3, which is scalable, durable, and cost-effective.It provides low latency access to the subset of the data that the researchersregularly require, as it is cached locally in the Storage Gateway appliance.It does not require the entire dataset to be accessed on a daily basis, as it isstored in Amazon S3 and can be retrieved on demand.It offers flexible data protection and recovery options, as it allows taking point-intimecopies of the volumes using AWS Backup, which are stored in AWS asAmazon EBS snapshots.Therefore, the solutions architect should recommend deploying an AWS Storage Gatewayvolume gateway with cached volumes with an Amazon S3 bucket as the target storage andmigrating the data to the Storage Gateway appliance.References:Volume Gateway | Amazon Web ServicesHow Volume Gateway works (architecture) - AWS Storage GatewayAWS Storage Volume Gateway - Cached volumes - Stack Overflow
Question # 353
A company runs a three-tier web application in a VPC across multiple Availability Zones. Amazon EC2 instances run in an Auto Scaling group for the application tier. The company needs to make an automated scaling plan that will analyze each resource's daily and weekly historical workload trends. The configuration must scale resources appropriately according to both the forecast and live changes in utilization. Which scaling strategy should a solutions architect recommend to meet these requirements?
A. Implement dynamic scaling with step scaling based on average CPU utilization from theEC2 instances. B. Enable predictive scaling to forecast and scale. Configure dynamic scaling with targettracking. C. Create an automated scheduled scaling action based on the traffic patterns of the webapplication. D. Set up a simple scaling policy. Increase the cooldown period based on the EC2 instancestartup time
Answer: B Explanation:This solution meets the requirements because it allows the company to use both predictivescaling and dynamic scaling to optimize the capacity of its Auto Scaling group. Predictivescaling uses machine learning to analyze historical data and forecast future traffic patterns.It then adjusts the desired capacity of the group in advance of the predicted changes.Dynamic scaling uses target tracking to maintain a specified metric (such as CPUutilization) at a target value. It scales the group in or out as needed to keep the metric close to the target. By using both scaling methods, the company can benefit from faster, simpler,and more accurate scaling that responds to both forecasted and live changes in utilization.References:Predictive scaling for Amazon EC2 Auto Scaling[Target tracking scaling policies for Amazon EC2 Auto Scaling
Question # 354
A company deployed a serverless application that uses Amazon DynamoDB as a database layer The application has experienced a large increase in users. The company wants to improve database response time from milliseconds to microseconds and to cache requests to the database. Which solution will meet these requirements with the LEAST operational overhead?
A. Use DynamoDB Accelerator (DAX). B. Migrate the database to Amazon Redshift. C. Migrate the database to Amazon RDS. D. Use Amazon ElastiCache for Redis.
Answer: A Explanation: DynamoDB Accelerator (DAX) is a fully managed, highly available cachingservice built for Amazon DynamoDB. DAX delivers up to a 10 times performanceimprovement—from milliseconds to microseconds—even at millions of requests persecond. DAX does all the heavy lifting required to add in-memory acceleration to yourDynamoDB tables, without requiring developers to manage cache invalidation, datapopulation, or cluster management. Now you can focus on building great applications foryour customers without worrying about performance at scale. You do not need to modifyapplication logic because DAX is compatible with existing DynamoDB API calls. Thissolution will meet the requirements with the least operational overhead, as it does notrequire any code development or manual intervention. References:1 provides an overview of Amazon DynamoDB Accelerator (DAX) and its benefits.2 explains how to use DAX with DynamoDB for in-memory acceleration.
Question # 355
An online video game company must maintain ultra-low latency for its game servers. The game servers run on Amazon EC2 instances. The company needs a solution that can handle millions of UDP internet traffic requests each second. Which solution will meet these requirements MOST cost-effectively?
A. Configure an Application Load Balancer with the required protocol and ports for theinternet traffic. Specify the EC2 instances as the targets. B. Configure a Gateway Load Balancer for the internet traffic. Specify the EC2 instances asthe targets. C. Configure a Network Load Balancer with the required protocol and ports for the internettraffic. Specify the EC2 instances as the targets. D. Launch an identical set of game servers on EC2 instances in separate AWS Regions. Route internet traffic to both sets of EC2 instances.
Answer: C Explanation: The most cost-effective solution for the online video game company is toconfigure a Network Load Balancer with the required protocol and ports for the internettraffic and specify the EC2 instances as the targets. This solution will enable the companyto handle millions of UDP requests per second with ultra-low latency and high performance.A Network Load Balancer is a type of Elastic Load Balancing that operates at theconnection level (Layer 4) and routes traffic to targets (EC2 instances, microservices, orcontainers) within Amazon VPC based on IP protocol data. A Network Load Balancer isideal for load balancing of both TCP and UDP traffic, as it is capable of handling millions ofrequests per second while maintaining high throughput at ultra-low latency. A NetworkLoad Balancer also preserves the source IP address of the clients to the back-endapplications, which can be useful for logging or security purposes1.
Question # 356
A company maintains an Amazon RDS database that maps users to cost centers. The company has accounts in an organization in AWS Organizations. The company needs a solution that will tag all resources that are created in a specific AWS account in the organization. The solution must tag each resource with the cost center ID of the user who created the resource. Which solution will meet these requirements?
A. Move the specific AWS account to a new organizational unit (OU) in Organizations fromthe management account. Create a service control policy (SCP) that requires all existingresources to have the correct cost center tag before the resources are created. Apply the SCP to the new OU. B. Create an AWS Lambda function to tag the resources after the Lambda function looksup the appropriate cost center from the RDS database. Configure an Amazon EventBridgerule that reacts to AWS CloudTrail events to invoke the Lambda function. C. Create an AWS CloudFormation stack to deploy an AWS Lambda function. Configurethe Lambda function to look up the appropriate cost center from the RDS database and totag resources. Create an Amazon EventBridge scheduled rule to invoke theCloudFormation stack. D. Create an AWS Lambda function to tag the resources with a default value. Configure anAmazon EventBridge rule that reacts to AWS CloudTrail events to invoke the Lambdafunction when a resource is missing the cost center tag.
Answer: B Explanation: AWS Lambda is a serverless compute service that lets you run code without provisioning or managing servers. Lambda can be used to tag resources with the costcenter ID of the user who created the resource, by querying the RDS database that mapsusers to cost centers. Amazon EventBridge is a serverless event bus service that enablesevent-driven architectures. EventBridge can be configured to react to AWS CloudTrailevents, which are recorded API calls made by or on behalf of the AWS account.EventBridge can invoke the Lambda function when a resource is created in the specificAWS account, passing the user identity and resource information as parameters. Thissolution will meet the requirements, as it enables automatic tagging of resources based onthe user and cost center mapping.References:1 provides an overview of AWS Lambda and its benefits.2 provides an overview of Amazon EventBridge and its benefits.3 explains the concept and benefits of AWS CloudTrail events.
Question # 357
A company is designing a tightly coupled high performance computing (HPC) environment in the AWS Cloud The company needs to include features that will optimize the HPC environment for networking and storage. Which combination of solutions will meet these requirements? (Select TWO )
A. Create an accelerator in AWS Global Accelerator. Configure custom routing for theaccelerator. B. Create an Amazon FSx for Lustre file system. Configure the file system with scratchstorage. C. Create an Amazon CloudFront distribution. Configure the viewer protocol policy to beHTTP and HTTPS. D. Launch Amazon EC2 instances. Attach an Elastic Fabric Adapter (EFA) to theinstances. E. Create an AWS Elastic Beanstalk deployment to manage the environment.
Answer: B,D Explanation: These two solutions will optimize the HPC environment for networking and storage. Amazon FSx for Lustre is a fully managed service that provides cost-effective,high-performance, scalable storage for compute workloads. It is built on the world’s mostpopular high-performance file system, Lustre, which is designed for applications thatrequire fast storage, such as HPC and machine learning. By configuring the file systemwith scratch storage, you can achieve sub-millisecond latencies, up to hundreds of GBs/sof throughput, and millions of IOPS. Scratch file systems are ideal for temporary storageand shorter-term processing of data. Data is not replicated and does not persist if a fileserver fails. For more information, see Amazon FSx for Lustre.Elastic Fabric Adapter (EFA) is a network interface for Amazon EC2 instances that enablescustomers to run applications requiring high levels of inter-node communications at scaleon AWS. Its custom-built operating system (OS) bypass hardware interface enhances theperformance of inter-instance communications, which is critical to scaling HPC andmachine learning applications. EFA provides a low-latency, low-jitter channel for interinstancecommunications, enabling your tightly-coupled HPC or distributed machinelearning applications to scale to thousands of cores. EFA uses libfabric interface andlibfabric APIs for communications, which are supported by most HPC programmingmodels. For more information, see Elastic Fabric Adapter. The other solutions are not suitable for optimizing the HPC environment for networking andstorage. AWS Global Accelerator is a networking service that helps you improve theavailability, performance, and security of your public applications by using the AWS globalnetwork. It provides two global static public IPs, deterministic routing, fast failover, and TCPtermination at the edge for your application endpoints. However, it does not support OSbypasscapabilities or high-performance file systems that are required for HPC andmachine learning applications. For more information, see AWS Global Accelerator.Amazon CloudFront is a content delivery network (CDN) service that securely deliversdata, videos, applications, and APIs to customers globally with low latency, high transferspeeds, all within a developer-friendly environment. CloudFront is integrated with AWSservices such as Amazon S3, Amazon EC2, AWS Elemental Media Services, AWS Shield,AWS WAF, and AWS Lambda@Edge. However, CloudFront is not designed for HPC andmachine learning applications that require high levels of inter-node communications andfast storage. For more information, see [Amazon CloudFront].AWS Elastic Beanstalk is an easy-to-use service for deploying and scaling webapplications and services developed with Java, .NET, PHP, Node.js, Python, Ruby, Go,and Docker on familiar servers such as Apache, Nginx, Passenger, and IIS. You cansimply upload your code and Elastic Beanstalk automatically handles the deployment, fromcapacity provisioning, load balancing, auto-scaling to application health monitoring.However, Elastic Beanstalk is not optimized for HPC and machine learning applicationsthat require OS-bypass capabilities and high-performance file systems. For moreinformation, see [AWS Elastic Beanstalk].References: Amazon FSx for Lustre, Elastic Fabric Adapter, AWS Global Accelerator,[Amazon CloudFront], [AWS Elastic Beanstalk].
Question # 358
A company is running a photo hosting service in the us-east-1 Region. The service enables users across multiple countries to upload and view photos. Some photos are heavily viewed for months, and others are viewed for less than a week. The application allows uploads of up to 20 MB for each photo. The service uses the photo metadata to determine which photos to display to each user. Which solution provides the appropriate user access MOST cost-effectively?
A. Store the photos in Amazon DynamoDB. Turn on DynamoDB Accelerator (DAX) tocache frequently viewed items. B. Store the photos in the Amazon S3 Intelligent-Tiering storage class. Store the photometadata and its S3 location in DynamoDB. C. Store the photos in the Amazon S3 Standard storage class. Set up an S3 Lifecyclepolicy to move photos older than 30 days to the S3 Standard-Infrequent Access (S3Standard-IA) storage class. Use the object tags to keep track of metadata. D. Store the photos in the Amazon S3 Glacier storage class. Set up an S3 Lifecycle policyto move photos older than 30 days to the S3 Glacier Deep Archive storage class. Store thephoto metadata and its S3 location in Amazon OpenSearch Service.
Answer: B Explanation: This solution provides the appropriate user access most cost-effectivelybecause it uses the Amazon S3 Intelligent-Tiering storage class, which automaticallyoptimizes storage costs by moving data to the most cost-effective access tier when access patterns change, without performance impact or operational overhead1. This storage classis ideal for data with unknown, changing, or unpredictable access patterns, such as photosthat are heavily viewed for months or less than a week. By storing the photo metadata andits S3 location in DynamoDB, the application can quickly query and retrieve the relevantphotos for each user. DynamoDB is a fast, scalable, and fully managed NoSQL databaseservice that supports key-value and document data models2.References: 1: Amazon S3 Intelligent-Tiering Storage Class | AWS3, Overview section2:Amazon DynamoDB - NoSQL Cloud Database Service4, Overview section.
Question # 359
A company is designing a new web application that will run on Amazon EC2 Instances. The application will use Amazon DynamoDB for backend data storage. The application traffic will be unpredictable. T company expects that the application read and write throughput to the database will be moderate to high. The company needs to scale in response to application traffic. Which DynamoDB table configuration will meet these requirements MOST cost-effectively?
A. Configure DynamoDB with provisioned read and write by using the DynamoDBStandard table class. Set DynamoDB auto scaling to a maximum defined capacity. B. Configure DynamoDB in on-demand mode by using the DynamoDB Standard tableclass. C. Configure DynamoDB with provisioned read and write by using the DynamoDBStandard Infrequent Access (DynamoDB Standard-IA) table class. Set DynamoDB autoscaling to a maximum defined capacity. D. Configure DynamoDB in on-demand mode by using the DynamoDB Standard InfrequentAccess (DynamoDB Standard-IA) table class.
Answer: B Explanation: The most cost-effective DynamoDB table configuration for the web application is to configure DynamoDB in on-demand mode by using the DynamoDBStandard table class. This configuration will allow the company to scale in response toapplication traffic and pay only for the read and write requests that the application performson the table.On-demand mode is a flexible billing option that can handle thousands of requests persecond without capacity planning. On-demand mode automatically adjusts the table’scapacity based on the incoming traffic, and charges only for the read and write requeststhat are actually performed. On-demand mode is suitable for applications withunpredictable or variable workloads, or applications that prefer the ease of paying for onlywhat they use1.The DynamoDB Standard table class is the default and recommended table class for mostworkloads. The DynamoDB Standard table class offers lower throughput costs than theDynamoDB Standard-Infrequent Access (DynamoDB Standard-IA) table class, and is morecost-effective for tables where throughput is the dominant cost. The DynamoDB Standardtable class also offers the same performance, durability, and availability as the DynamoDBStandard-IA table class2. The other options are not correct because they are either not cost-effective or not suitablefor the use case. Configuring DynamoDB with provisioned read and write by using theDynamoDB Standard table class, and setting DynamoDB auto scaling to a maximumdefined capacity is not correct because this configuration requires manual estimation andmanagement of the table’s capacity, which adds complexity and cost to the solution.Provisioned mode is a billing option that requires users to specify the amount of read andwrite capacity units for their tables, and charges for the reserved capacity regardless ofusage. Provisioned mode is suitable for applications with predictable or stable workloads,or applications that require finer-grained control over their capacity settings1. ConfiguringDynamoDB with provisioned read and write by using the DynamoDB Standard-InfrequentAccess (DynamoDB Standard-IA) table class, and setting DynamoDB auto scaling to amaximum defined capacity is not correct because this configuration is not cost-effective fortables with moderate to high throughput. The DynamoDB Standard-IA table class offerslower storage costs than the DynamoDB Standard table class, but higher throughput costs.The DynamoDB Standard-IA table class is optimized for tables where storage is thedominant cost, such as tables that store infrequently accessed data2. ConfiguringDynamoDB in on-demand mode by using the DynamoDB Standard-Infrequent Access(DynamoDB Standard-IA) table class is not correct because this configuration is not costeffectivefor tables with moderate to high throughput. As mentioned above, the DynamoDBStandard-IA table class has higher throughput costs than the DynamoDB Standard tableclass, which can offset the savings from lower storage costs.References:Table classes - Amazon DynamoDBRead/write capacity mode - Amazon DynamoDB
Question # 360
A company's web application that is hosted in the AWS Cloud recently increased in popularity. The web application currently exists on a single Amazon EC2 instance in a single public subnet. The web application has not been able to meet the demand of the increased web traffic. The company needs a solution that will provide high availability and scalability to meet the increased user demand without rewriting the web application. Which combination of steps will meet these requirements? (Select TWO.)
A. Replace the EC2 instance with a larger compute optimized instance. B. Configure Amazon EC2 Auto Scaling with multiple Availability Zones in private subnets. C. Configure a NAT gateway in a public subnet to handle web requests. D. Replace the EC2 instance with a larger memory optimized instance. E. Configure an Application Load Balancer in a public subnet to distribute web traffic
Answer: B,E Explanation:These two steps will meet the requirements because they will provide high availability andscalability for the web application without rewriting it. Amazon EC2 Auto Scaling allows youto automatically adjust the number of EC2 instances in response to changes in demand. Byconfiguring Auto Scaling with multiple Availability Zones in private subnets, you can ensurethat your web application is distributed across isolated and fault-tolerant locations, and thatyour instances are not directly exposed to the internet. An Application Load Balanceroperates at the application layer and distributes incoming web traffic across multipletargets, such as EC2 instances, containers, or Lambda functions. By configuring anApplication Load Balancer in a public subnet, you can enable your web application tohandle requests from the internet and route them to the appropriate targets in the privatesubnets.References:What is Amazon EC2 Auto Scaling?What is an Application Load Balancer?