Google Professional Machine Learning Engineer Exam Practice Questions (P. 4)
- Full Access (339 questions)
- Six months of Premium Access
- Access to one million comments
- Seamless ChatGPT Integration
- Ability to download PDF files
- Anki Flashcard files for revision
- No Captcha & No AdSense
- Advanced Exam Configuration
Question #31
You need to train a computer vision model that predicts the type of government ID present in a given image using a GPU-powered virtual machine on Compute
Engine. You use the following parameters:
✑ Optimizer: SGD
✑ Image shape = 224ֳ—224
✑ Batch size = 64
✑ Epochs = 10
✑ Verbose =2
During training you encounter the following error: ResourceExhaustedError: Out Of Memory (OOM) when allocating tensor. What should you do?
Engine. You use the following parameters:
✑ Optimizer: SGD
✑ Image shape = 224ֳ—224
✑ Batch size = 64
✑ Epochs = 10
✑ Verbose =2
During training you encounter the following error: ResourceExhaustedError: Out Of Memory (OOM) when allocating tensor. What should you do?
- AChange the optimizer.
- BReduce the batch size.Most Voted
- CChange the learning rate.
- DReduce the image shape.
Correct Answer:
B
Reference:
https://github.com/tensorflow/tensorflow/issues/136
B
Reference:
https://github.com/tensorflow/tensorflow/issues/136
send
light_mode
delete
Question #32
You developed an ML model with AI Platform, and you want to move it to production. You serve a few thousand queries per second and are experiencing latency issues. Incoming requests are served by a load balancer that distributes them across multiple Kubeflow CPU-only pods running on Google Kubernetes Engine
(GKE). Your goal is to improve the serving latency without changing the underlying infrastructure. What should you do?
(GKE). Your goal is to improve the serving latency without changing the underlying infrastructure. What should you do?
- ASignificantly increase the max_batch_size TensorFlow Serving parameter.
- BSwitch to the tensorflow-model-server-universal version of TensorFlow Serving.
- CSignificantly increase the max_enqueued_batches TensorFlow Serving parameter.
- DRecompile TensorFlow Serving using the source to support CPU-specific optimizations. Instruct GKE to choose an appropriate baseline minimum CPU platform for serving nodes.Most Voted
Correct Answer:
D
D
send
light_mode
delete
Question #33
You have a demand forecasting pipeline in production that uses Dataflow to preprocess raw data prior to model training and prediction. During preprocessing, you employ Z-score normalization on data stored in BigQuery and write it back to BigQuery. New training data is added every week. You want to make the process more efficient by minimizing computation time and manual intervention. What should you do?
- ANormalize the data using Google Kubernetes Engine.
- BTranslate the normalization algorithm into SQL for use with BigQuery.Most Voted
- CUse the normalizer_fn argument in TensorFlow's Feature Column API.
- DNormalize the data with Apache Spark using the Dataproc connector for BigQuery.
Correct Answer:
B
B
send
light_mode
delete
Question #34
You need to design a customized deep neural network in Keras that will predict customer purchases based on their purchase history. You want to explore model performance using multiple model architectures, store training data, and be able to compare the evaluation metrics in the same dashboard. What should you do?
- ACreate multiple models using AutoML Tables.
- BAutomate multiple training runs using Cloud Composer.
- CRun multiple training jobs on AI Platform with similar job names.
- DCreate an experiment in Kubeflow Pipelines to organize multiple runs.Most Voted
Correct Answer:
C
C
send
light_mode
delete
Question #35
You are developing a Kubeflow pipeline on Google Kubernetes Engine. The first step in the pipeline is to issue a query against BigQuery. You plan to use the results of that query as the input to the next step in your pipeline. You want to achieve this in the easiest way possible. What should you do?
- AUse the BigQuery console to execute your query, and then save the query results into a new BigQuery table.
- BWrite a Python script that uses the BigQuery API to execute queries against BigQuery. Execute this script as the first step in your Kubeflow pipeline.
- CUse the Kubeflow Pipelines domain-specific language to create a custom component that uses the Python BigQuery client library to execute queries.
- DLocate the Kubeflow Pipelines repository on GitHub. Find the BigQuery Query Component, copy that component's URL, and use it to load the component into your pipeline. Use the component to execute queries against BigQuery.Most Voted
Correct Answer:
A
A
send
light_mode
delete
Question #36
You are building a model to predict daily temperatures. You split the data randomly and then transformed the training and test datasets. Temperature data for model training is uploaded hourly. During testing, your model performed with 97% accuracy; however, after deploying to production, the model's accuracy dropped to 66%. How can you make your production model more accurate?
- ANormalize the data for the training, and test datasets as two separate steps.
- BSplit the training and test data based on time rather than a random split to avoid leakage.Most Voted
- CAdd more data to your test set to ensure that you have a fair distribution and sample for testing.
- DApply data transformations before splitting, and cross-validate to make sure that the transformations are applied to both the training and test sets.
Correct Answer:
D
D
send
light_mode
delete
Question #37
You are developing models to classify customer support emails. You created models with TensorFlow Estimators using small datasets on your on-premises system, but you now need to train the models using large datasets to ensure high performance. You will port your models to Google Cloud and want to minimize code refactoring and infrastructure overhead for easier migration from on-prem to cloud. What should you do?
- AUse AI Platform for distributed training.Most Voted
- BCreate a cluster on Dataproc for training.
- CCreate a Managed Instance Group with autoscaling.
- DUse Kubeflow Pipelines to train on a Google Kubernetes Engine cluster.
Correct Answer:
C
C
send
light_mode
delete
Question #38
You have trained a text classification model in TensorFlow using AI Platform. You want to use the trained model for batch predictions on text data stored in
BigQuery while minimizing computational overhead. What should you do?
BigQuery while minimizing computational overhead. What should you do?
- AExport the model to BigQuery ML.Most Voted
- BDeploy and version the model on AI Platform.
- CUse Dataflow with the SavedModel to read the data from BigQuery.
- DSubmit a batch prediction job on AI Platform that points to the model location in Cloud Storage.
Correct Answer:
A
A
send
light_mode
delete
Question #39
You work with a data engineering team that has developed a pipeline to clean your dataset and save it in a Cloud Storage bucket. You have created an ML model and want to use the data to refresh your model as soon as new data is available. As part of your CI/CD workflow, you want to automatically run a Kubeflow
Pipelines training job on Google Kubernetes Engine (GKE). How should you architect this workflow?
Pipelines training job on Google Kubernetes Engine (GKE). How should you architect this workflow?
- AConfigure your pipeline with Dataflow, which saves the files in Cloud Storage. After the file is saved, start the training job on a GKE cluster.
- BUse App Engine to create a lightweight python client that continuously polls Cloud Storage for new files. As soon as a file arrives, initiate the training job.
- CConfigure a Cloud Storage trigger to send a message to a Pub/Sub topic when a new file is available in a storage bucket. Use a Pub/Sub-triggered Cloud Function to start the training job on a GKE cluster.Most Voted
- DUse Cloud Scheduler to schedule jobs at a regular interval. For the first step of the job, check the timestamp of objects in your Cloud Storage bucket. If there are no new files since the last run, abort the job.
Correct Answer:
C
C
send
light_mode
delete
Question #40
You have a functioning end-to-end ML pipeline that involves tuning the hyperparameters of your ML model using AI Platform, and then using the best-tuned parameters for training. Hypertuning is taking longer than expected and is delaying the downstream processes. You want to speed up the tuning job without significantly compromising its effectiveness. Which actions should you take? (Choose two.)
- ADecrease the number of parallel trials.
- BDecrease the range of floating-point values.
- CSet the early stopping parameter to TRUE.Most Voted
- DChange the search algorithm from Bayesian search to random search.Most Voted
- EDecrease the maximum number of trials during subsequent training phases.
Correct Answer:
BD
Reference:
https://cloud.google.com/ai-platform/training/docs/hyperparameter-tuning-overview
BD
Reference:
https://cloud.google.com/ai-platform/training/docs/hyperparameter-tuning-overview
send
light_mode
delete
All Pages