AWS Certified Data Analytics ā Specialty (DAS-C01) Exam Simulator #5
Simulator Summary
0 of 100 Questions completed
Questions:
Information
You have already completed the simulator before. Hence you can not start it again.
Simulator is loading…
You must sign in or sign up to start the simulator.
You must first complete the following:
Results
Results
0 of 100 Questions answered correctly
Your time:
Time has elapsed
You have reached 0 of 0 point(s), (0)
Earned Point(s): 0 of 0, (0)
0 Essay(s) Pending (Possible Point(s): 0)
Categories
- Not categorized 0%
-
Unfortunately, you didn’t pass the practice exam, but hey, you have unlimited access.š
Practise makes you perfect! š -
Congratulations! š„³
You have passed the practice exam successfully! You are one step closer to pass the real exam!
We hope to see you again on another certification path.āļø
Good luck with the exam! Stay strong.š
- 1
- 2
- 3
- 4
- 5
- 6
- 7
- 8
- 9
- 10
- 11
- 12
- 13
- 14
- 15
- 16
- 17
- 18
- 19
- 20
- 21
- 22
- 23
- 24
- 25
- 26
- 27
- 28
- 29
- 30
- 31
- 32
- 33
- 34
- 35
- 36
- 37
- 38
- 39
- 40
- 41
- 42
- 43
- 44
- 45
- 46
- 47
- 48
- 49
- 50
- 51
- 52
- 53
- 54
- 55
- 56
- 57
- 58
- 59
- 60
- 61
- 62
- 63
- 64
- 65
- 66
- 67
- 68
- 69
- 70
- 71
- 72
- 73
- 74
- 75
- 76
- 77
- 78
- 79
- 80
- 81
- 82
- 83
- 84
- 85
- 86
- 87
- 88
- 89
- 90
- 91
- 92
- 93
- 94
- 95
- 96
- 97
- 98
- 99
- 100
- Current
- Review
- Answered
- Correct
- Incorrect
-
Question 1 of 100
1. Question
Which ingestion service is best suited for low-volume, high-latency data with unpredictable processing requirements?
CorrectIncorrect -
Question 2 of 100
2. Question
A data collection system is required to handle a high volume of data with high throughput capabilities. Which AWS service is best suited for this scenario?
CorrectIncorrect -
Question 3 of 100
3. Question
A company needs to collect and process data from a large number of remote sensors that transmit small amounts of data frequently. Which AWS service is best suited for this scenario?
CorrectIncorrect -
Question 4 of 100
4. Question
Which collection solution should you use when dealing with source data systems that have intermittent connectivity?
CorrectIncorrect -
Question 5 of 100
5. Question
Which collection solution should you use when dealing with source data systems that are located outside of the AWS cloud and do not support an AWS SDK or CLI?
CorrectIncorrect -
Question 6 of 100
6. Question
Which collection solution should you use to capture changes in data at the source, ensuring that the data retains its order and format?
CorrectIncorrect -
Question 7 of 100
7. Question
Which collection solution should you use to capture changes in data at the source, ensuring that the data is compressed and stored in a specific format?
CorrectIncorrect -
Question 8 of 100
8. Question
Which of the following collection systems should be used to collect log data with a fixed format and size?
CorrectIncorrect -
Question 9 of 100
9. Question
Which of the following compression formats can be used for Amazon S3 data storage?
CorrectIncorrect -
Question 10 of 100
10. Question
Which of the following processing methods ensures that each record in a data stream is processed only once, while maintaining low latency?
CorrectIncorrect -
Question 11 of 100
11. Question
Which of the following collection systems is suitable for collecting log data with out-of-order delivery and duplicates?
CorrectIncorrect -
Question 12 of 100
12. Question
Which of the following collection systems allows for in-flight data transformation and filtering during the collection process?
CorrectIncorrect -
Question 13 of 100
13. Question
Which of the following transformation methods is suitable for converting data from a proprietary format to a standard format during the collection process?
CorrectIncorrect -
Question 14 of 100
14. Question
A data analytics company needs to process and analyze large sets of data from various sources. The data is gathered and sent in batches, with each batch containing millions of records. Which of the following collection systems is best suited for this requirement?
CorrectIncorrect -
Question 15 of 100
15. Question
A data analytics team needs to ingest large volumes of data from various sources while ensuring high fault tolerance and low latency. Which ingestion service should they choose?
CorrectIncorrect -
Question 16 of 100
16. Question
What is the best storage solution for structured data that requires low latency and high throughput?
CorrectIncorrect -
Question 17 of 100
17. Question
What is the best storage solution for semi-structured data that requires cost-effective and scalable storage?
CorrectIncorrect -
Question 18 of 100
18. Question
You are tasked with designing a data analytics solution that requires the fastest possible query response times. Which AWS storage solution would be the most appropriate to meet this requirement?
CorrectIncorrect -
Question 19 of 100
19. Question
You are designing a data analytics solution for a company that needs to store and query large amounts of structured data in a data warehouse. Which AWS storage solution would be the most appropriate to meet this requirement?
CorrectIncorrect -
Question 20 of 100
20. Question
You are working on a data analytics project that requires frequent updates to the schema. Which AWS service would be the most appropriate to handle schema evolution requirements?
CorrectIncorrect -
Question 21 of 100
21. Question
You are working on a data analytics project that requires a high degree of flexibility in the data schema. Which AWS service would be the most appropriate to handle this requirement?
CorrectIncorrect -
Question 22 of 100
22. Question
Which storage format should you use to optimize performance for a data analysis task that requires querying a subset of data from a large dataset stored in Amazon S3?
CorrectIncorrect -
Question 23 of 100
23. Question
Which compression strategy should you use to reduce the storage size and improve the query performance of a large dataset stored in Amazon S3 in Parquet format?
CorrectIncorrect -
Question 24 of 100
24. Question
Which encoding strategy should you use to improve the query performance of a large dataset stored in Amazon S3 in ORC format?
CorrectIncorrect -
Question 25 of 100
25. Question
You have a large dataset stored in Amazon S3 that you need to analyze using Amazon Athena. The dataset contains millions of records and you need to optimize the query performance. Which data sorting and distribution strategy should you use?
CorrectIncorrect -
Question 26 of 100
26. Question
You have a large dataset stored in Amazon Redshift that you need to analyze using SQL queries. The dataset contains hundreds of millions of records and you need to optimize the query performance. Which storage layout should you use?
CorrectIncorrect -
Question 27 of 100
27. Question
Which of the following data distribution methods would have the highest performance with the lowest cost for a large-scale data analytics workload on AWS?
CorrectIncorrect -
Question 28 of 100
28. Question
What is a potential cost and performance implication of using a columnar data format like Apache Parquet instead of a row-based format like CSV for a data analytics workload on AWS?
CorrectIncorrect -
Question 29 of 100
29. Question
Which of the following is an appropriate partitioning scheme for data-optimized analysis when dealing with a large dataset that has a time component?
CorrectIncorrect -
Question 30 of 100
30. Question
Which of the following data formats is most suitable for a data lake that needs to support a wide variety of query types and allows for schema-on-read?
CorrectIncorrect -
Question 31 of 100
31. Question
Which strategy is best suited for a company that primarily works with data generated by Internet of Things (IoT) devices and requires high write throughput?
CorrectIncorrect -
Question 32 of 100
32. Question
A company needs to store its archived data for regulatory compliance reasons. The data is infrequently accessed, and the company requires a cost-effective solution. Which solution would best meet the company’s requirements?
CorrectIncorrect -
Question 33 of 100
33. Question
A company has a large dataset that is used for analytical purposes. The data is constantly growing, and the company needs to ensure that only the most recent data is available for analysis. Which AWS service would be best suited to meet the company’s requirements?
CorrectIncorrect -
Question 34 of 100
34. Question
A company has a large dataset that contains sensitive information. The company needs to ensure that the data is stored securely and is not accessible to unauthorized users. Which AWS service would be best suited to meet the company’s requirements?
CorrectIncorrect -
Question 35 of 100
35. Question
Which mechanism can be used to discover new data sources in a data lake architecture?
CorrectIncorrect -
Question 36 of 100
36. Question
Which service can be used to automatically detect and capture changes in a relational database and make them available in a data lake?
CorrectIncorrect -
Question 37 of 100
37. Question
Which AWS service can be used to create and manage a centralized metadata repository for all data assets in an organization?
CorrectIncorrect -
Question 38 of 100
38. Question
Which mechanism can be used to update the metadata in a data catalog automatically when changes are made to the data sources?
CorrectIncorrect -
Question 39 of 100
39. Question
Which mechanism can be used to search and retrieve data catalogs and metadata in AWS Glue?
CorrectIncorrect -
Question 40 of 100
40. Question
What is the recommended mechanism for searching and retrieving data catalogs and metadata in Amazon EMR?
CorrectIncorrect -
Question 41 of 100
41. Question
Which AWS service allows you to classify data by applying custom labels to objects?
CorrectIncorrect -
Question 42 of 100
42. Question
Which AWS service allows you to tag data with custom metadata attributes to describe its contents?
CorrectIncorrect -
Question 43 of 100
43. Question
A company has a large amount of historical data that is rarely accessed but needs to be kept for compliance reasons. Which AWS storage solution would be the best option to store this data?
CorrectIncorrect -
Question 44 of 100
44. Question
A company is analyzing a large dataset of customer reviews for a product. The dataset is stored in Amazon S3 and contains information such as the review text, rating, and reviewer ID. The company wants to optimize query performance by using a data format that can compress the data and minimize the amount of storage space used. Which data format should they use?
CorrectIncorrect -
Question 45 of 100
45. Question
A gaming company needs to store and query data from a high-performance analytics platform that requires low latency and high throughput. The data is structured and requires frequent updates. What storage solution is most appropriate to address query latency requirements?
CorrectIncorrect -
Question 46 of 100
46. Question
Which of the following AWS services is best suited for running repeatable ETL workflows using Apache Spark and Apache Hive?
CorrectIncorrect -
Question 47 of 100
47. Question
A data processing job in an AWS Glue ETL script has failed due to a network connectivity issue with the data source. Which AWS service can be used to recover from this processing failure?
CorrectIncorrect -
Question 48 of 100
48. Question
Which AWS service can be used to automate the recovery of an AWS Glue ETL job in case of a processing failure?
CorrectIncorrect -
Question 49 of 100
49. Question
A data analyst needs to ensure auditing and traceability for a data processing job that runs on an Amazon EMR cluster. Which AWS service can be used to deploy logging and monitoring solutions for this purpose?
CorrectIncorrect -
Question 50 of 100
50. Question
A data processing job that runs on an Amazon EC2 instance is failing due to insufficient disk space. Which AWS service can be used to deploy logging and monitoring solutions to enable auditing and traceability for this processing job?
CorrectIncorrect -
Question 51 of 100
51. Question
Which technique can help address concurrency needs in a data transformation and preparation solution when using Amazon EMR?
CorrectIncorrect -
Question 52 of 100
52. Question
You are designing a data analytics solution that involves processing data in real-time. You need to apply appropriate ELT techniques for this workload. What is a suitable ELT technique to use in this scenario?
CorrectIncorrect -
Question 53 of 100
53. Question
You are designing a data analytics solution that requires concurrent access to the same dataset. You need to implement a mechanism that ensures that multiple users can access the dataset simultaneously without conflicts. What is a suitable concurrency technique to use in this scenario?
CorrectIncorrect -
Question 54 of 100
54. Question
You are designing a data processing solution that involves performing transformations on a large dataset stored in Amazon S3. The transformations involve filtering, aggregating, and sorting the data. You want to optimize the solution for cost efficiency. Which AWS service should you use?
CorrectIncorrect -
Question 55 of 100
55. Question
You are building a data processing solution that involves ingesting and transforming data from a variety of sources. Some of the data is only needed for a short period of time, while other data needs to be stored long-term for analysis. Which AWS service should you use to optimize cost efficiency?
CorrectIncorrect -
Question 56 of 100
56. Question
Which AWS service can be used to aggregate and enrich data for downstream consumption?
CorrectIncorrect -
Question 57 of 100
57. Question
Which AWS service can be used to automate the process of creating and running machine learning workflows?
CorrectIncorrect -
Question 58 of 100
58. Question
A company is using AWS Glue to automate their data processing workflows. They want to ensure that their workflows automatically recover from any processing failures. Which of the following methods can be used to identify and recover from processing failures in AWS Glue?
CorrectIncorrect -
Question 59 of 100
59. Question
A data analyst is building an automated data processing workflow using AWS Data Pipeline. The workflow consists of several activities, including data ingestion, data transformation, and data loading. The analyst wants to ensure that the workflow is resilient to failures and can recover automatically. Which of the following techniques can the analyst use to recover from processing failures in AWS Data Pipeline?
CorrectIncorrect -
Question 60 of 100
60. Question
Which of the following is a feature of Amazon CloudWatch Logs that can be used to monitor, store, and access log files from Amazon Elastic Compute Cloud (Amazon EC2) instances?
CorrectIncorrect -
Question 61 of 100
61. Question
Which of the following tools can be used to aggregate and enrich data for downstream consumption in a batch processing data analytics solution?
CorrectIncorrect -
Question 62 of 100
62. Question
Which mechanism can be used to replicate Amazon S3 objects between regions for disaster recovery purposes?
CorrectIncorrect -
Question 63 of 100
63. Question
A data analytics team is planning to process large amounts of data in a scalable and cost-effective way. They need a solution that can handle both batch and real-time processing, and provides automatic scaling and fault tolerance. Which AWS service would best fit their requirements?
CorrectIncorrect -
Question 64 of 100
64. Question
A company needs to process data in real-time and respond to events immediately. Which of the following services would be the best choice for this requirement?
CorrectIncorrect -
Question 65 of 100
65. Question
Which of the following logging and monitoring solutions would be the most appropriate for auditing and traceability of data processing solutions on AWS?
CorrectIncorrect -
Question 66 of 100
66. Question
You have created a sales dashboard that displays monthly sales data. The data is generated once a day by a batch process that runs at midnight. Which data refresh schedule is most appropriate for this scenario?
CorrectIncorrect -
Question 67 of 100
67. Question
You have created a dashboard that provides real-time updates on customer transactions for a financial institution. The data for this dashboard is being generated continuously by various sources. Which data refresh schedule is most appropriate for this scenario?
CorrectIncorrect -
Question 68 of 100
68. Question
Which method is appropriate for delivering data to a remote team of data scientists who need to collaborate on a machine learning project using Jupyter notebooks?
CorrectIncorrect -
Question 69 of 100
69. Question
Which method is appropriate for delivering data to a team of data analysts who need to collaborate on a project using Jupyter notebooks?
CorrectIncorrect -
Question 70 of 100
70. Question
Which data delivery method is appropriate for a marketing campaign report that requires real-time updates and access to data via email for a remote team?
CorrectIncorrect -
Question 71 of 100
71. Question
Which data delivery method is appropriate for a financial report that needs to be sent to multiple stakeholders via email?
CorrectIncorrect -
Question 72 of 100
72. Question
Which data delivery method is appropriate for a data visualization application that requires offline access to data and analysis on mobile devices for executives?
CorrectIncorrect -
Question 73 of 100
73. Question
Which data delivery method is appropriate for a business intelligence application that requires real-time access to data and analysis on mobile devices for field sales representatives?
CorrectIncorrect -
Question 74 of 100
74. Question
In a business intelligence solution, a company needs to deliver real-time dashboards for their customers to view sales performance. Which method of data delivery is the most appropriate for this scenario?
CorrectIncorrect -
Question 75 of 100
75. Question
A company has a mobile application that generates a lot of data, and they want to track the user behavior in real-time. They want a solution that can provide real-time insights and allow them to interact with the data. Which AWS service would be the best fit for this scenario?
CorrectIncorrect -
Question 76 of 100
76. Question
A company has developed a new mobile application and wants to track user behavior for different features of the application. The company wants to use an API to integrate the data with their existing analytics tool. Which AWS service would be the best fit for this scenario?
CorrectIncorrect -
Question 77 of 100
77. Question
You have a dataset of employee performance ratings and want to display the distribution of ratings for each department. Which data visualization solution should you use to display this tabular data?
CorrectIncorrect -
Question 78 of 100
78. Question
You have a large dataset of customer transactions and want to display the total revenue generated by each product category. Which data visualization solution should you use to display this tabular data?
CorrectIncorrect -
Question 79 of 100
79. Question
You are analyzing customer feedback on a new product feature and want to display the overall satisfaction rating. Which data visualization solution should you use to display this KPI?
CorrectIncorrect -
Question 80 of 100
80. Question
You are analyzing the revenue generated by different product lines over the past year. Which data visualization solution should you use to display the Key Performance Indicators (KPIs) for this analysis?
CorrectIncorrect -
Question 81 of 100
81. Question
You are analyzing the performance of a website and want to display the number of pageviews by hour of the day. Which data visualization solution should you use to display this data?
CorrectIncorrect -
Question 82 of 100
82. Question
You are analyzing customer data to determine the most popular product category sold during the holiday season. Which data visualization solution should you use to display this data?
CorrectIncorrect -
Question 83 of 100
83. Question
A company wants to analyze their customers’ behavior and preferences using historical data to improve their marketing strategies. Which type of analysis solution should they use?
CorrectIncorrect -
Question 84 of 100
84. Question
A company wants to monitor the performance of their website in real-time and quickly detect anomalies to improve their customers’ experience. Which type of analysis solution should they use?
CorrectIncorrect -
Question 85 of 100
85. Question
Which AWS service can be used to ensure compliance with the General Data Protection Regulation (GDPR) requirements for personal data protection?
CorrectIncorrect -
Question 86 of 100
86. Question
Which of the following AWS services can help meet compliance requirements for data residency?
CorrectIncorrect -
Question 87 of 100
87. Question
Which of the following is a benefit of enabling AWS CloudTrail?
CorrectIncorrect -
Question 88 of 100
88. Question
Which of the following AWS services allow for access and audit logging?
CorrectIncorrect -
Question 89 of 100
89. Question
Your organization has recently implemented a new cloud-based data analytics platform, and you need to ensure that you are meeting all necessary data governance and compliance requirements. Which of the following is the most important consideration when determining your data governance and compliance requirements?
CorrectIncorrect -
Question 90 of 100
90. Question
As a data analyst, you are working with sensitive financial data and need to ensure that your data governance and compliance requirements are met. What is the first step in determining your data governance and compliance requirements?
CorrectIncorrect -
Question 91 of 100
91. Question
What is a best practice for key rotation in AWS?
CorrectIncorrect -
Question 92 of 100
92. Question
What is a best practice for managing secrets in AWS?
CorrectIncorrect -
Question 93 of 100
93. Question
Which of the following is an example of data masking?
CorrectIncorrect -
Question 94 of 100
94. Question
Which of the following is an example of data obfuscation?
CorrectIncorrect -
Question 95 of 100
95. Question
Which of the following AWS services provides automatic server-side encryption for data at rest by default?
CorrectIncorrect -
Question 96 of 100
96. Question
Which of the following is an example of server-side encryption for data at rest in an AWS environment?
CorrectIncorrect -
Question 97 of 100
97. Question
Which of the following techniques should be used to mask data in a development environment that contains personally identifiable information (PII)?
CorrectIncorrect -
Question 98 of 100
98. Question
Which of the following data types require encryption in transit and at rest for compliance with data protection regulations?
CorrectIncorrect -
Question 99 of 100
99. Question
Which of the following statements accurately describes role-based access control?
CorrectIncorrect -
Question 100 of 100
100. Question
Which of the following is a feature of Amazon S3 server-side encryption?
Correct Answer: B It encrypts data at rest with AES-256 encryption.
Explanation:Amazon S3 server-side encryption uses AES-256 encryption to encrypt data at rest. It does not use client-side encryption with AWS KMS-managed keys, nor does it encrypt data in transit using SSL/TLS. Although it is possible to use a user-defined encryption key, it is not a feature of Amazon S3 server-side encryption.
Option A is incorrect because Amazon S3 server-side encryption does not use client-side encryption with AWS KMS-managed keys.
Option C is incorrect because Amazon S3 server-side encryption does not encrypt data in transit using SSL/TLS.
Option D is incorrect because while it is possible to use a user-defined encryption key, it is not a feature of Amazon S3 server-side encryption.CorrectIncorrect