1. A media content company has a streaming playback application. The company wants to collect and analyze the data to provide near-real-time feedback on playback issues. The company needs to consume this data and return results within 30 seconds according to the service-level agreement (SLA). The company needs the consumer to identify playback issues, such as quality during a specified timeframe. The data will be emitted as JSON and may change schemas over time.Which solution will allow the company to collect data for processing while meeting these requirements?
A) Send the data to Amazon Kinesis Data Firehose with delivery to Amazon S3. Configure an S3 event trigger an AWS Lambda function to process the data. The Lambda function will consume the data and process it to identify potential playback issues. Persist the raw data to Amazon S3. B) Send the data to Amazon Managed Streaming for Kafka and configure an Amazon Kinesis Analytics for Java application as the consumer. The application will consume the data and process it to identify potential playback issues. Persist the raw data to Amazon DynamoDB. C) Send the data to Amazon Kinesis Data Firehose with delivery to Amazon S3. Configure Amazon S3 to trigger an event for AWS Lambda to process. The Lambda function will consume the data and process it to identify potential playback issues. Persist the raw data to Amazon DynamoDB. D) Send the data to Amazon Kinesis Data Streams and configure an Amazon Kinesis Analytics for Java application as the consumer. The application will consume the data and process it to identify potential playback issues. Persist the raw data to Amazon S3.
2. An ecommerce company stores customer purchase data in Amazon RDS. The company wants a solution to store and analyze historical data. The most recent 6 months of data will be queried frequently for analytics workloads. This data is several terabytes large. Once a month, historical data for the last 5 years must be accessible and will be joined with the more recent data. The company wants to optimize performance and cost.Which storage solution will meet these requirements?
A) Create a read replica of the RDS database to store the most recent 6 months of data. Copy the historical data into Amazon S3. Create an AWS Glue Data Catalog of the data in Amazon S3 and Amazon RDS. Run historical queries using Amazon Athena. B) Use an ETL tool to incrementally load the most recent 6 months of data into an Amazon Redshift cluster. Run more frequent queries against this cluster. Create a read replica of the RDS database to run queries on the historical data. C) Incrementally copy data from Amazon RDS to Amazon S3. Create an AWS Glue Data Catalog of the data in Amazon S3. Use Amazon Athena to query the data. D) Incrementally copy data from Amazon RDS to Amazon S3. Load and store the most recent 6 months of data in Amazon Redshift. Configure an Amazon Redshift Spectrum table to connect to all historical data.
3. A company leverages Amazon Athena for ad-hoc queries against data stored in Amazon S3. The company wants to implement additional controls to separate query execution and query history among users, teams, or applications running in the same AWS account to comply with internal security policies.Which solution meets these requirements?
A) Create an S3 bucket for each given use case, create an S3 bucket policy that grants permissions to appropriate individual IAM users. and apply the S3 bucket policy to the S3 bucket. B) Create an Athena workgroup for each given use case, apply tags to the workgroup, and create an IAM policy using the tags to apply appropriate permissions to the workgroup. C) Create an IAM role for each given use case, assign appropriate permissions to the role for the given use case, and add the role to associate the role with Athena. D) Create an AWS Glue Data Catalog resource policy for each given use case that grants permissions to appropriate individual IAM users, and apply the resource policy to the specific tables used by Athena.
4. A company wants to use an automatic machine learning (ML) Random Cut Forest (RCF) algorithm to visualize complex real-word scenarios, such as detecting seasonality and trends, excluding outers, and imputing missing values.The team working on this project is non-technical and is looking for an out-of-the-box solution that will require the LEAST amount of management overhead.Which solution will meet these requirements?
A) Use an AWS Glue ML transform to create a forecast and then use Amazon QuickSight to visualize the data. B) Use Amazon QuickSight to visualize the data and then use ML-powered forecasting to forecast the key business metrics. C) Use a pre-build ML AMI from the AWS Marketplace to create forecasts and then use Amazon QuickSight to visualize the data. D) Use calculated fields to create a new forecast and then use Amazon QuickSight to visualize the data.
5. A retail company's data analytics team recently created multiple product sales analysis dashboards for the average selling price per product using AmazonQuickSight. The dashboards were created from .csv files uploaded to Amazon S3. The team is now planning to share the dashboards with the respective external product owners by creating individual users in Amazon QuickSight. For compliance and governance reasons, restricting access is a key requirement. The product owners should view only their respective product analysis in the dashboard reports.Which approach should the data analytics team take to allow product owners to view only their products in the dashboard?
A) Separate the data by product and use S3 bucket policies for authorization. B) Separate the data by product and use IAM policies for authorization. C) Create a manifest file with row-level security. D) Create dataset rules with row-level security.
1. Right Answer: B Explanation: SLA within 30 secs rules out options with Firehose. And DynamoDB works fine for schema changes on JSON
2. Right Answer: D Explanation: Historical data should be stored on s3 and joined with most recent data from Redshift using spectrum (external tables)
3. Right Answer: B Explanation: Amazon Athena Workgroups - A new resource type that can be used to separate query execution and query history between Users, Teams, or Applications running under the same AWS account Referencehttps://aws.amazon.com/about-aws/whats-new/2019/02/athena_workgroups/
4. Right Answer: B Explanation: Referenceshttps://docs.aws.amazon.com/quicksight/latest/user/making-data-driven-decisions-with-ml-in-quicksight.htmlhttps://docs.aws.amazon.com/quicksight/latest/user/what-is-random-cut-forest.html
5. Right Answer: D Explanation: Referencehttps://docs.aws.amazon.com/quicksight/latest/user/restrict-access-to-a-data-set-using-row-level-security.html
Leave a comment