Amazon AWS Certified Data Engineer - Associate DEA-C01 Exam Practice Questions (P. 4)
- Full Access (208 questions)
- Six months of Premium Access
- Access to one million comments
- Seamless ChatGPT Integration
- Ability to download PDF files
- Anki Flashcard files for revision
- No Captcha & No AdSense
- Advanced Exam Configuration
Question #16
A company uses an on-premises Microsoft SQL Server database to store financial transaction data. The company migrates the transaction data from the on-premises database to AWS at the end of each month. The company has noticed that the cost to migrate data from the on-premises database to an Amazon RDS for SQL Server database has increased recently.
The company requires a cost-effective solution to migrate the data to AWS. The solution must cause minimal downtown for the applications that access the database.
Which AWS service should the company use to meet these requirements?
The company requires a cost-effective solution to migrate the data to AWS. The solution must cause minimal downtown for the applications that access the database.
Which AWS service should the company use to meet these requirements?
- AAWS Lambda
- BAWS Database Migration Service (AWS DMS)Most Voted
- CAWS Direct Connect
- DAWS DataSync
Correct Answer:
D
D
send
light_mode
delete
Question #17
A data engineer is building a data pipeline on AWS by using AWS Glue extract, transform, and load (ETL) jobs. The data engineer needs to process data from Amazon RDS and MongoDB, perform transformations, and load the transformed data into Amazon Redshift for analytics. The data updates must occur every hour.
Which combination of tasks will meet these requirements with the LEAST operational overhead? (Choose two.)
Which combination of tasks will meet these requirements with the LEAST operational overhead? (Choose two.)
- AConfigure AWS Glue triggers to run the ETL jobs every hour.Most Voted
- BUse AWS Glue DataBrew to clean and prepare the data for analytics.
- CUse AWS Lambda functions to schedule and run the ETL jobs every hour.
- DUse AWS Glue connections to establish connectivity between the data sources and Amazon Redshift.Most Voted
- EUse the Redshift Data API to load transformed data into Amazon Redshift.
Correct Answer:
BC
BC
send
light_mode
delete
Question #18
A company uses an Amazon Redshift cluster that runs on RA3 nodes. The company wants to scale read and write capacity to meet demand. A data engineer needs to identify a solution that will turn on concurrency scaling.
Which solution will meet this requirement?
Which solution will meet this requirement?
- ATurn on concurrency scaling in workload management (WLM) for Redshift Serverless workgroups.
- BTurn on concurrency scaling at the workload management (WLM) queue level in the Redshift cluster.Most Voted
- CTurn on concurrency scaling in the settings during the creation of any new Redshift cluster.
- DTurn on concurrency scaling for the daily usage quota for the Redshift cluster.
Correct Answer:
D
D
send
light_mode
delete
Question #19
A data engineer must orchestrate a series of Amazon Athena queries that will run every day. Each query can run for more than 15 minutes.
Which combination of steps will meet these requirements MOST cost-effectively? (Choose two.)
Which combination of steps will meet these requirements MOST cost-effectively? (Choose two.)
- AUse an AWS Lambda function and the Athena Boto3 client start_query_execution API call to invoke the Athena queries programmatically.Most Voted
- BCreate an AWS Step Functions workflow and add two states. Add the first state before the Lambda function. Configure the second state as a Wait state to periodically check whether the Athena query has finished using the Athena Boto3 get_query_execution API call. Configure the workflow to invoke the next query when the current query has finished running.Most Voted
- CUse an AWS Glue Python shell job and the Athena Boto3 client start_query_execution API call to invoke the Athena queries programmatically.
- DUse an AWS Glue Python shell script to run a sleep timer that checks every 5 minutes to determine whether the current Athena query has finished running successfully. Configure the Python shell script to invoke the next query when the current query has finished running.
- EUse Amazon Managed Workflows for Apache Airflow (Amazon MWAA) to orchestrate the Athena queries in AWS Batch.
Correct Answer:
CD
CD
send
light_mode
delete
Question #20
A company is migrating on-premises workloads to AWS. The company wants to reduce overall operational overhead. The company also wants to explore serverless options.
The company's current workloads use Apache Pig, Apache Oozie, Apache Spark, Apache Hbase, and Apache Flink. The on-premises workloads process petabytes of data in seconds. The company must maintain similar or better performance after the migration to AWS.
Which extract, transform, and load (ETL) service will meet these requirements?
The company's current workloads use Apache Pig, Apache Oozie, Apache Spark, Apache Hbase, and Apache Flink. The on-premises workloads process petabytes of data in seconds. The company must maintain similar or better performance after the migration to AWS.
Which extract, transform, and load (ETL) service will meet these requirements?
send
light_mode
delete
All Pages