Decide Fast & Get 50% Flat Discount on This SPECIAL OFFER | Limited Time Offer - Ends In COUPON CODE: SAVE50%

Free Amazon DAS-C01 Questions

    1. Page: 1/32
      Total 157 Questions
    Question No 1
    A financial services company needs to aggregate daily stock trade data from the exchanges into a data store. The company requires that data be streamed directly into the data store, but also occasionally allows data to be modified using SQL. The solution should integrate complex, analytic queries running with minimal latency. The solution must provide a business intelligence dashboard that enables viewing of the top contributors to anomalies in stock prices. Which solution meets the company’s requirements?
    Choose the Choices:


    Question No 2
    A financial company hosts a data lake in Amazon S3 and a data warehouse on an Amazon Redshift cluster. The company uses Amazon QuickSight to build dashboards and wants to secure access from its on - premises Active Directory to Amazon QuickSight. How should the data be secured?
    Choose the Choices:


    Question No 3
    A real estate company has a mission - critical application using Apache HBase in Amazon EMR. Amazon EMR is configured with a single master node. The company has over 5 TB of data stored on an Hadoop Distributed File System (HDFS). The company wants a cost - effective solution to make its HBase data highly available. Which architectural pattern meets company’s requirements?
    Choose the Choices:


    Question No 4
    A software company hosts an application on AWS, and new features are released weekly. As part of the application testing process, a solution must be developed that analyzes logs from each Amazon EC2 instance to ensure that the application is working as expected after each deployment. The collection and analysis solution should be highly available with the ability to display new information with minimal delays. Which method should the company use to collect and analyze the logs?
    Choose the Choices:


    Question No 5
    A data analyst is using AWS Glue to organize, cleanse, validate, and format a 200 GB dataset. The data analyst triggered the job to run with the Standard worker type. After 3 hours, the AWS Glue job status is still RUNNING. Logs from the job run show no error codes. The data analyst wants to improve the job execution time without overprovisioning. Which actions should the data analyst take?
    Choose the Choices:



    1. Page: 1/32
      Total 157 Questions