Amazon AWS Certified Big Data - Specialty Exam Practice Questions (P. 1)
- Full Access (85 questions)
- Six months of Premium Access
- Access to one million comments
- Seamless ChatGPT Integration
- Ability to download PDF files
- Anki Flashcard files for revision
- No Captcha & No AdSense
- Advanced Exam Configuration
Question #1
Redshift.
What is the most efficient architecture strategy for this purpose?
- ATransform the unstructured data using Amazon EMR and generate CSV data. COPY the CSV data into the analysis schema within Redshift.
- BLoad the unstructured data into Redshift, and use string parsing functions to extract structured data for inserting into the analysis schema.
- CWhen the data is saved to Amazon S3, use S3 Event Notifications and AWS Lambda to transform the file contents. Insert the data into the analysis schema on Redshift.
- DNormalize the data using an AWS Marketplace ETL tool, persist the results to Amazon S3, and use AWS Lambda to INSERT the data into Redshift.
A

Hi! Do you need help with this question ?
- Why isn't the A the right answer?
- Traducir la pregunta al español
Contributor get free access to an augmented ChatGPT 4 trained with the latest IT Questions.
Question #2
Which AWS service strategy is best for this use case?
- ACopy the data into Amazon ElastiCache to perform text analysis on the in-memory data and export the results of the model into Amazon Machine Learning.
- BUse Amazon EMR to parallelize the text analysis tasks across the cluster using a streaming program step.
- CUse Amazon Elasticsearch Service to store the text and then use the Python Elasticsearch Client to run analysis against the text index.
- DInitiate a Python job from AWS Data Pipeline to run directly against the Amazon S3 text files.
C
Reference: https://aws.amazon.com/blogs/database/indexing-metadata-in-amazon-elasticsearch-service- using-aws-lambda-and-python/

Hi! Do you need help with this question ?
- Why isn't the A the right answer?
- Traducir la pregunta al español
Contributor get free access to an augmented ChatGPT 4 trained with the latest IT Questions.
Question #3
How should this control mapping be achieved using AWS?
- ARequest AWS third-party audit reports and/or the AWS quality addendum and map the AWS responsibilities to the controls that must be provided.
- BRequest data center Temporary Auditor access to an AWS data center to verify the control mapping.
- CRequest relevant SLAs and security guidelines for Amazon DynamoDB and define these guidelines within the applications architecture to map to the control framework.
- DRequest Amazon DynamoDB system architecture designs to determine how to map the AWS responsibilities to the control that must be provided.
A

Hi! Do you need help with this question ?
- Why isn't the A the right answer?
- Traducir la pregunta al español
Contributor get free access to an augmented ChatGPT 4 trained with the latest IT Questions.
Question #4
In which three circumstances would choosing Key-based distribution be most appropriate? (Select three.)
- AWhen the administrator needs to optimize a large, slowly changing dimension table.
- BWhen the administrator needs to reduce cross-node traffic.
- CWhen the administrator needs to optimize the fact table for parity with the number of slices.
- DWhen the administrator needs to balance data distribution and collocation data.
- EWhen the administrator needs to take advantage of data locality on a local node for joins and aggregates.
ACD

Hi! Do you need help with this question ?
- Why isn't the A the right answer?
- Traducir la pregunta al español
Contributor get free access to an augmented ChatGPT 4 trained with the latest IT Questions.
Question #5
Which option allows Company A to do clustering in the AWS Cloud and meet the legal requirement of maintaining personal data in-country?
- AAnonymize the personal data portions of the dataset and transfer the data files into Amazon S3 in the AWS region. Have the EMR cluster read the dataset using EMRFS.
- BEstablish a Direct Connect link between the on-premises system and the AWS region to reduce latency. Have the EMR cluster read the data directly from the on-premises storage system over Direct Connect.
- CEncrypt the data files according to encryption standards of Country X and store them on AWS region in Amazon S3. Have the EMR cluster read the dataset using EMRFS.
- DUse AWS Import/Export Snowball device to securely transfer the data to the AWS region and copy the files onto an EBS volume. Have the EMR cluster read the dataset using EMRFS.
B

Hi! Do you need help with this question ?
- Why isn't the A the right answer?
- Traducir la pregunta al español
Contributor get free access to an augmented ChatGPT 4 trained with the latest IT Questions.
All Pages