0% found this document useful (0 votes)
175 views120 pages

Google - Professional Machine Learning Engineer.v2024 10 23.q109

The document outlines various questions and answers related to the Google Professional Machine Learning Engineer certification exam. It covers topics such as building and orchestrating training pipelines, automating unit tests, evaluating model performance, and deploying models using Vertex AI. The answers provide insights into best practices and recommended approaches for machine learning tasks in Google Cloud.

Uploaded by

asif ali
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
175 views120 pages

Google - Professional Machine Learning Engineer.v2024 10 23.q109

The document outlines various questions and answers related to the Google Professional Machine Learning Engineer certification exam. It covers topics such as building and orchestrating training pipelines, automating unit tests, evaluating model performance, and deploying models using Vertex AI. The answers provide insights into best practices and recommended approaches for machine learning tasks in Google Cloud.

Uploaded by

asif ali
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 120

Google.

Professional-Machine-Learning-
Engineer.v2024-10-23.q109

Exam Code: Professional-Machine-Learning-Engineer


Exam Name: Google Professional Machine Learning Engineer
Certification Provider: Google
Free Question Number: 109
Version: v2024-10-23
# of views: 104
# of Questions views: 1090
https://www.freecram.net/torrent/Google.Professional-Machine-Learning-
Engineer.v2024-10-23.q109.html

NEW QUESTION: 1
You are an ML engineer responsible for designing and implementing training pipelines for ML
models. You need to create an end-to-end training pipeline for a TensorFlow model. The
TensorFlow model will be trained on several terabytes of structured data. You need the pipeline
to include data quality checks before training and model quality checks after training but prior to
deployment. You want to minimize development time and the need for infrastructure
maintenance. How should you build and orchestrate your training pipeline?
A. Create the pipeline using Kubeflow Pipelines domain-specific language (DSL) and predefined
Google Cloud components. Orchestrate the pipeline using Vertex AI Pipelines.
B. Create the pipeline using TensorFlow Extended (TFX) and standard TFX components.
Orchestrate the pipeline using Vertex AI Pipelines.
C. Create the pipeline using Kubeflow Pipelines domain-specific language (DSL) and predefined
Google Cloud components. Orchestrate the pipeline using Kubeflow Pipelines deployed on
Google Kubernetes Engine.
D. Create the pipeline using TensorFlow Extended (TFX) and standard TFX components.
Orchestrate the pipeline using Kubeflow Pipelines deployed on Google Kubernetes Engine.
Answer: (SHOW ANSWER)
The best option for creating and orchestrating an end-to-end training pipeline for a TensorFlow
model is to use TensorFlow Extended (TFX) and standard TFX components, and deploy the
pipeline to Vertex AI Pipelines. TFX is an end-to-end platform for deploying production ML
pipelines, which consists of several built-in components that cover the entire ML lifecycle, from
data ingestion and validation, to model training and evaluation, to model deployment and
monitoring. TFX also supports custom components and integrations with other Google Cloud
services, such as BigQuery, Dataflow, and Cloud Storage. Vertex AI Pipelines is a fully managed
service that allows you to run TFX pipelines on Google Cloud, without having to worry about
infrastructure provisioning, scaling, or maintenance. Vertex AI Pipelines also provides a user-
friendly interface to monitor and manage your pipelines, as well as tools to track and compare
experiments. The other options are not as suitable for creating and orchestrating an end-to-end
training pipeline for a TensorFlow model, because:
* Creating the pipeline using Kubeflow Pipelines domain-specific language (DSL) and predefined
Google Cloud components would require more development time and effort, as Kubeflow
Pipelines DSL is not as expressive or compatible with TensorFlow as TFX. Predefined Google
Cloud components might not cover all the stages of the ML lifecycle, and might not be optimized
for TensorFlow models.
* Orchestrating the pipeline using Kubeflow Pipelines deployed on Google Kubernetes Engine
would require more infrastructure maintenance, as Kubeflow Pipelines is not a fully managed
service, and you would have to provision and manage your own Kubernetes cluster. This would
also incur more costs, as
* you would have to pay for the cluster resources, regardless of the pipeline usage. References:
* TFX | ML Production Pipelines | TensorFlow
* Vertex AI Pipelines | Google Cloud
* Kubeflow Pipelines | Google Cloud
* Google Cloud launches machine learning engineer certification
* Google Professional Machine Learning Engineer Certification
* Professional ML Engineer Exam Guide

NEW QUESTION: 2
You have written unit tests for a Kubeflow Pipeline that require custom libraries. You want to
automate the execution of unit tests with each new push to your development branch in Cloud
Source Repositories. What should you do?
A. Write a script that sequentially performs the push to your development branch and executes
the unit tests on Cloud Run
B. Using Cloud Build, set an automated trigger to execute the unit tests when changes are
pushed to your development branch.
C. Set up a Cloud Logging sink to a Pub/Sub topic that captures interactions with Cloud Source
Repositories Configure a Pub/Sub trigger for Cloud Run, and execute the unit tests on Cloud
Run.
D. Set up a Cloud Logging sink to a Pub/Sub topic that captures interactions with Cloud Source
Repositories. Execute the unit tests using a Cloud Function that is triggered when messages are
sent to the Pub/Sub topic
Answer: (SHOW ANSWER)
Cloud Build is a service that executes your builds on Google Cloud Platform infrastructure. Cloud
Build can import source code from Cloud Source Repositories, Cloud Storage, GitHub, or
Bitbucket, execute a build to your specifications, and produce artifacts such as Docker containers
or Java archives1 Cloud Build allows you to set up automated triggers that start a build when
changes are pushed to a source code repository. You can configure triggers to filter the changes
based on the branch, tag, or file path2 To automate the execution of unit tests for a Kubeflow
Pipeline that require custom libraries, you can use Cloud Build to set an automated trigger to
execute the unit tests when changes are pushed to your development branch in Cloud Source
Repositories. You can specify the steps of the build in a YAML or JSON file, such as installing the
custom libraries, running the unit tests, and reporting the results. You can also use Cloud Build to
build and deploy the Kubeflow Pipeline components if the unit tests pass3 The other options are
not recommended or feasible. Writing a script that sequentially performs the push to your
development branch and executes the unit tests on Cloud Run is not a good practice, as it does
not leverage the benefits of Cloud Build and its integration with Cloud Source Repositories.
Setting up a Cloud Logging sink to a Pub/Sub topic that captures interactions with Cloud Source
Repositories and using a Pub/Sub trigger for Cloud Run or Cloud Function to execute the unit
tests is unnecessarily complex and inefficient, as it adds extra steps and latency to the process.
Cloud Run and Cloud Function are also not designed for executing unit tests, as they have
limitations on the memory, CPU, and execution time45 References: 1: Cloud Build overview 2:
Creating and managing build triggers 3: Building and deploying Kubeflow Pipelines using Cloud
Build 4: Cloud Run documentation 5: Cloud Functions documentation

NEW QUESTION: 3
Your company manages a video sharing website where users can watch and upload videos. You
need to create an ML model to predict which newly uploaded videos will be the most popular so
that those videos can be prioritized on your company's website. Which result should you use to
determine whether the model is successful?
A. The model predicts videos as popular if the user who uploads them has over 10,000 likes.
B. The model predicts 97.5% of the most popular clickbait videos measured by number of clicks.
C. The model predicts 95% of the most popular videos measured by watch time within 30 days of
being uploaded.
D. The Pearson correlation coefficient between the log-transformed number of views after 7 days
and 30 days after publication is equal to 0.
Answer: (SHOW ANSWER)
In this scenario, the goal is to create an ML model to predict which newly uploaded videos will be
the most popular on a video sharing website. The result that should be used to determine whether
the model is successful is the one that best aligns with the business objective and the evaluation
metric. Option C is the correct answer because it defines the most popular videos as the ones
that have the highest watch time within
30 days of being uploaded, and it sets a high accuracy threshold of 95% for the model prediction.
Option C: The model predicts 95% of the most popular videos measured by watch time within 30
days of being uploaded. This option is the best result for the scenario because it reflects the
business objective and the evaluation metric. The business objective is to prioritize the videos
that will attract and retain the most viewers on the website. The watch time is a good indicator of
the viewer engagement and satisfaction, as it measures how long the viewers watch the videos.
The 30-day window is a reasonable time frame to capture the popularity trend of the videos, as it
accounts for the initial interest and the viral potential of the videos. The
95% accuracy threshold is a high standard for the model prediction, as it means that the model
can correctly identify 95 out of 100 of the most popular videos based on the watch time metric.
Option A: The model predicts videos as popular if the user who uploads them has over 10,000
likes. This option is not a good result for the scenario because it does not reflect the business
objective or the evaluation metric. The business objective is to prioritize the videos that will be the
most popular on the website, not the users who upload them. The number of likes that a user has
is not a good indicator of the popularity of their videos, as it does not measure the viewer
engagement or satisfaction with the videos. Moreover, this option does not specify a time frame
or an accuracy threshold for the model prediction, making it vague and unreliable.
Option B: The model predicts 97.5% of the most popular clickbait videos measured by number of
clicks. This option is not a good result for the scenario because it does not reflect the business
objective or the evaluation metric. The business objective is to prioritize the videos that will be the
most popular on the website, not the videos that have the most misleading or sensational titles or
thumbnails. The number of clicks that a video has is not a good indicator of the popularity of the
video, as it does not measure the viewer engagement or satisfaction with the video content.
Moreover, this option only focuses on the clickbait videos, which may not represent the majority
or the diversity of the videos on the website.
Option D: The Pearson correlation coefficient between the log-transformed number of views after
7 days and
30 days after publication is equal to 0. This option is not a good result for the scenario because it
does not reflect the business objective or the evaluation metric. The business objective is to
prioritize the videos that will be the most popular on the website, not the videos that have the
most consistent or inconsistent number of views over time. The Pearson correlation coefficient is
a metric that measures the linear relationship between two variables, not the popularity of the
videos. A correlation coefficient of 0 means that there is no linear relationship between the log-
transformed number of views after 7 days and 30 days, which does not indicate whether the
videos are popular or not. Moreover, this option does not specify a threshold or a target value for
the correlation coefficient, making it meaningless and irrelevant.

NEW QUESTION: 4
You work for an organization that operates a streaming music service. You have a custom
production model that is serving a "next song" recommendation based on a user's recent listening
history. Your model is deployed on a Vertex Al endpoint. You recently retrained the same model
by using fresh data. The model received positive test results offline. You now want to test the new
model in production while minimizing complexity. What should you do?
A. Create a new Vertex Al endpoint for the new model and deploy the new model to that new
endpoint Build a service to randomly send 5% of production traffic to the new endpoint Monitor
end-user metrics such as listening time If end-user metrics improve between models over time
gradually increase the percentage of production traffic sent to the new endpoint.
B. Capture incoming prediction requests in BigQuery Create an experiment in Vertex Al
Experiments Run batch predictions for both models using the captured data Use the user's
selected song to compare the models performance side by side If the new models performance
metrics are better than the previous model deploy the new model to production.
C. Deploy the new model to the existing Vertex Al endpoint Use traffic splitting to send 5% of
production traffic to the new model Monitor end-user metrics, such as listening time If end-user
metrics improve between models over time, gradually increase the percentage of production
traffic sent to the new model.
D. Configure a model monitoring job for the existing Vertex Al endpoint. Configure the monitoring
job to detect prediction drift, and set a threshold for alerts Update the model on the endpoint from
the previous model to the new model If you receive an alert of prediction drift, revert to the
previous model.
Answer: (SHOW ANSWER)
Traffic splitting is a feature of Vertex AI that allows you to distribute the prediction requests among
multiple models or model versions within the same endpoint. You can specify the percentage of
traffic that each model or model version receives, and change it at any time. Traffic splitting can
help you test the new model in production without creating a new endpoint or a separate service.
You can deploy the new model to the existing Vertex AI endpoint, and use traffic splitting to send
5% of production traffic to the new model. You can monitor the end-user metrics, such as
listening time, to compare the performance of the new model and the previous model. If the end-
user metrics improve between models over time, you can gradually increase the percentage of
production traffic sent to the new model. This solution can help you test the new model in
production while minimizing complexity and cost. References:
* Traffic splitting | Vertex AI
* Deploying models to endpoints | Vertex AI

NEW QUESTION: 5
You need to develop a custom TensorRow model that will be used for online predictions. The
training data is stored in BigQuery. You need to apply instance-level data transformations to the
data for model training and serving. You want to use the same preprocessing routine during
model training and serving. How should you configure the preprocessing routine?
A. Create a BigQuery script to preprocess the data, and write the result to another BigQuery
table.
B. Create a pipeline in Vertex Al Pipelines to read the data from BigQuery and preprocess it using
a custom preprocessing component.
C. Create a preprocessing function that reads and transforms the data from BigQuery Create a
Vertex Al custom prediction routine that calls the preprocessing function at serving time.
D. Create an Apache Beam pipeline to read the data from BigQuery and preprocess it by using
TensorFlow Transform and Dataflow.
Answer: (SHOW ANSWER)
According to the official exam guide1, one of the skills assessed in the exam is to "design, build,
and productionalize ML models to solve business challenges using Google Cloud technologies".
TensorFlow Transform2 is a library for preprocessing data with TensorFlow. TensorFlow
Transform enables you to define and execute distributed pre-processing or feature engineering
functions on large data sets, and then export the same functions as a TensorFlow graph for re-
use during training or serving. TensorFlow Transform can handle both instance-level and full-pass
data transformations. Apache Beam3 is an open source framework for building scalable and
portable data pipelines. Apache Beam supports both batch and streaming data processing.
Dataflow4 is a fully managed service for running Apache Beam pipelines on Google Cloud.
Dataflow handles the provisioning and management of the compute resources, as well as the
optimization and execution of the pipelines. Therefore, option D is the best way to configure the
preprocessing routine for the given use case, as it allows you to use the same preprocessing
logic during model training and serving, and leverage the scalability and performance of Dataflow.
The other options are not relevant or optimal for this scenario. References:
* Professional ML Engineer Exam Guide
* TensorFlow Transform
* Apache Beam
* Dataflow
* Google Professional Machine Learning Certification Exam 2023
* Latest Google Professional Machine Learning Engineer Actual Free Exam Questions

NEW QUESTION: 6
You are working on a prototype of a text classification model in a managed Vertex AI Workbench
notebook.
You want to quickly experiment with tokenizing text by using a Natural Language Toolkit (NLTK)
library.
How should you add the library to your Jupyter kernel?
A. Install the NLTK library from a terminal by using the pip install nltk command.
B. Write a custom Dataflow job that uses NLTK to tokenize your text and saves the output to
Cloud Storage.
C. Create a new Vertex Al Workbench notebook with a custom image that includes the NLTK
library.
D. Install the NLTK library from a Jupyter cell by using the! pip install nltk -user command.
Answer: (SHOW ANSWER)
NLTK is a Python library that provides a set of tools for natural language processing, such as
tokenization, stemming, tagging, parsing, and sentiment analysis. Tokenization is a process of
breaking a text into smaller units, such as words or sentences. You can use NLTK to quickly
experiment with tokenizing text in a managed Vertex AI Workbench notebook. A Vertex AI
Workbench notebook is a web-based interactive environment that allows you to write and execute
Python code on Google Cloud. You can install the NLTK library from a Jupyter cell by using the !
pip install nltk --user command. This command uses the pip package manager to install the NLTK
library for the current user. By installing the NLTK library from a Jupyter cell, you can avoid the
hassle of opening a terminal or creating a custom image for your notebook. References:
* NLTK documentation
* Vertex AI Workbench documentation
* Preparing for Google Cloud Certification: Machine Learning Engineer Professional Certificate

NEW QUESTION: 7
You are training a deep learning model for semantic image segmentation with reduced training
time. While using a Deep Learning VM Image, you receive the following error: The resource
'projects/deeplearning-platforn/zones/europe-west4-c/acceleratorTypes/nvidia-tesla-k80' was not
found. What should you do?
A. Ensure that you have GPU quota in the selected region.
B. Ensure that the required GPU is available in the selected region.
C. Ensure that you have preemptible GPU quota in the selected region.
D. Ensure that the selected GPU has enough GPU memory for the workload.
Answer: (SHOW ANSWER)
The error message indicates that the selected GPU type (nvidia-tesla-k80) is not available in the
selected region (europe-west4-c). This can happen when the GPU type is not supported in the
region, or when the GPU quota is exhausted in the region. To avoid this error, you should ensure
that the required GPU is available in the selected region before creating a Deep Learning VM
Image. You can use the following steps to check the GPU availability and quota:
* To check the GPU availability, you can use the gcloud compute accelerator-types list command
with the --filter flag to specify the GPU type and the region. For example, to check the availability
of nvidia-tesla-k80 in europe-west4-c, you can run:
gcloud compute accelerator-types list --filter="name=nvidia-tesla-k80 AND zone:europe-west4-c"
* If the command returns an empty result, it means that the GPU type is not supported in the
region. You can either choose a different GPU type or a different region that supports the GPU
type. You can use the same command without the --filter flag to list all the available GPU types
and regions. For example, to list all the available GPU types in europe-west4-c, you can run:
gcloud compute accelerator-types list --filter="zone:europe-west4-c"
* To check the GPU quota, you can use the gcloud compute regions describe command with the
--format flag to specify the region and the quota metric. For example, to check the quota for
nvidia-tesla-k80 in europe-west4-c, you can run:
gcloud compute regions describe europe-west4-c --format="value(quotas.NVIDIA_K80_GPUS)"
* If the command returns a value of 0, it means that the GPU quota is exhausted in the region.
You can either request more quota from Google Cloud or choose a different region that has
enough quota for the GPU type.
References:
* Troubleshooting | Deep Learning VM Images | Google Cloud
* Checking GPU availability
* Checking GPU quota
NEW QUESTION: 8
You work for a rapidly growing social media company. Your team builds TensorFlow
recommender models in an on-premises CPU cluster. The data contains billions of historical user
events and 100 000 categorical features. You notice that as the data increases the model training
time increases. You plan to move the models to Google Cloud You want to use the most scalable
approach that also minimizes training time. What should you do?
A. Deploy the training jobs by using TPU VMs with TPUv3 Pod slices, and use the
TPUEmbedding API.
B. Deploy the training jobs in an autoscaling Google Kubernetes Engine cluster with CPUs
C. Deploy a matrix factorization model training job by using BigQuery ML.
D. Deploy the training jobs by using Compute Engine instances with A100 GPUs and use the t f.
nn. embedding_lookup API.
Answer: (SHOW ANSWER)
TPU VMs with TPUv3 Pod slices are the most scalable and performant option for training large-
scale recommender models on Google Cloud. TPUv3 Pods can provide up to 2048 cores and 32
TB of memory, and can process billions of examples and features in minutes. The
TPUEmbedding API is designed to efficiently handle large-scale categorical features and
embeddings, and can reduce the memory footprint and communication overhead of the model.
The other options are either less scalable (B and C) or less efficient (D) for this use case.

NEW QUESTION: 9
You work at a mobile gaming startup that creates online multiplayer games Recently, your
company observed an increase in players cheating in the games, leading to a loss of revenue and
a poor user experience. You built a binary classification model to determine whether a player
cheated after a completed game session, and then send a message to other downstream
systems to ban the player that cheated Your model has performed well during testing, and you
now need to deploy the model to production You want your serving solution to provide immediate
classifications after a completed game session to avoid further loss of revenue. What should you
do?
A. Import the model into Vertex Al Model Registry. Use the Vertex Batch Prediction service to run
batch inference jobs.
B. Save the model files in a Cloud Storage Bucket Create a Cloud Function to read the model
files and make online inference requests on the Cloud Function.
C. Save the model files in a VM Load the model files each time there is a prediction request and
run an inference job on the VM.
D. Import the model into Vertex Al Model Registry Create a Vertex Al endpoint that hosts the
model and make online inference requests.
Answer: D (LEAVE A REPLY)
Online inference is a process where you send a single or a small number of prediction requests to
a model and get immediate responses1. Online inference is suitable for scenarios where you
need timely predictions, such as detecting cheating in online games. Online inference requires
that the model is deployed to an endpoint, which is a resource that provides a service URL for
prediction requests2.
Vertex AI Model Registry is a central repository where you can manage the lifecycle of your ML
models3. You can import models from various sources, such as custom models or AutoML
models, and assign them to different versions and aliases3. You can also deploy models to
endpoints, which are resources that provide a service URL for online prediction2.
By importing the model into Vertex AI Model Registry, you can leverage the Vertex AI features to
monitor and update the model3. You can use Vertex AI Experiments to track and compare the
metrics of different model versions, such as accuracy, precision, recall, and AUC. You can also
use Vertex AI Explainable AI to generate feature attributions that show how much each input
feature contributed to the model's prediction.
By creating a Vertex AI endpoint that hosts the model, you can use the Vertex AI Prediction
service to serve online inference requests2. Vertex AI Prediction provides various benefits, such
as scalability, reliability, security, and logging2. You can use the Vertex AI API or the Google
Cloud console to send online inference requests to the endpoint and get immediate
classifications4.
Therefore, the best option for your scenario is to import the model into Vertex AI Model Registry,
create a Vertex AI endpoint that hosts the model, and make online inference requests.
The other options are not suitable for your scenario, because they either do not provide
immediate classifications, such as using batch prediction or loading the model files each time, or
they do not use Vertex AI Prediction, which would require more development and maintenance
effort, such as creating a Cloud Function or a VM.
References:
* Online versus batch prediction | Vertex AI | Google Cloud
* Deploy a model to an endpoint | Vertex AI | Google Cloud
* Introduction to Vertex AI Model Registry | Google Cloud
* Get online predictions | Vertex AI | Google Cloud

NEW QUESTION: 10

A. Delegate feature engineering to BigQuery and remove it from the pipeline.


B. Add a GPU to the model training step.
C. Enable caching in all the steps of the Kubeflow pipeline.
D. Comment out the part of the pipeline that you are not currently updating.
Answer: (SHOW ANSWER)
Kubeflow Pipelines allows for efficient use of compute resources through parallel task execution
and caching, which eliminates redundant executions1. By enabling caching in all the steps of the
Kubeflow pipeline, you can avoid re-running the same steps when you execute the pipeline
multiple times. This can significantly speed up the pipeline execution and reduce your
development time without incurring additional costs

NEW QUESTION: 11
You have deployed a model on Vertex AI for real-time inference. During an online prediction
request, you get an "Out of Memory" error. What should you do?
A. Use batch prediction mode instead of online mode.
B. Send the request again with a smaller batch of instances.
C. Use base64 to encode your data before using it for prediction.
D. Apply for a quota increase for the number of prediction requests.
Answer: (SHOW ANSWER)
* Option A is incorrect because using batch prediction mode instead of online mode does not
solve the
"Out of Memory" error, but rather changes the latency and throughput of the prediction service.
Batch prediction mode is suitable for large-scale, asynchronous, and non-urgent predictions,
while online prediction mode is suitable for low-latency, synchronous, and real-time predictions1.
* Option B is correct because sending the request again with a smaller batch of instances can
reduce the memory consumption of the prediction service and avoid the "Out of Memory" error.
The batch size is the number of instances that are processed together in one request. A smaller
batch size means less data to load into memory at once2.
* Option C is incorrect because using base64 to encode your data before using it for prediction
does not reduce the memory consumption of the prediction service, but rather increases it.
Base64 encoding is a way of representing binary data as ASCII characters, which increases the
size of the data by about
33%3. Base64 encoding is only required for certain data types, such as images and audio, that
cannot be represented as JSON or CSV4.
* Option D is incorrect because applying for a quota increase for the number of prediction
requests does not solve the "Out of Memory" error, but rather increases the number of requests
that can be sent to the prediction service per day. Quotas are limits on the usage of Google Cloud
resources, such as CPU, memory, disk, and network5. Quotas do not affect the performance of
the prediction service, but rather the availability and cost of the service.
References:
* Choosing between online and batch prediction
* Online prediction input data
* Base64 encoding
* Preparing data for prediction
* Quotas and limits
NEW QUESTION: 12
You are developing a mode! to detect fraudulent credit card transactions. You need to prioritize
detection because missing even one fraudulent transaction could severely impact the credit card
holder. You used AutoML to tram a model on users' profile information and credit card transaction
data. After training the initial model, you notice that the model is failing to detect many fraudulent
transactions. How should you adjust the training parameters in AutoML to improve model
performance?
Choose 2 answers
A. Increase the score threshold.
B. Decrease the score threshold.
C. Add more positive examples to the training set.
D. Add more negative examples to the training set.
E. Reduce the maximum number of node hours for training.
Answer: (SHOW ANSWER)
The best options for adjusting the training parameters in AutoML to improve model performance
are to decrease the score threshold and add more positive examples to the training set. These
options can help increase the detection rate of fraudulent transactions, which is the priority for this
use case. The score threshold is a parameter that determines the minimum probability score that
a prediction must have to be classified as positive. Decreasing the score threshold can increase
the recall of the model, which is the proportion of actual positive cases that are correctly
identified. Increasing the recall can help reduce the number of false negatives, which are
fraudulent transactions that are missed by the model. However, decreasing the score threshold
can also decrease the precision of the model, which is the proportion of positive predictions that
are actually correct. Decreasing the precision can increase the number of false positives, which
are legitimate transactions that are flagged as fraudulent by the model. Therefore, there is a
trade-off between recall and precision, and the optimal score threshold depends on the business
objective and the cost of errors1.
Adding more positive examples to the training set can help balance the data distribution and
improve the model performance. Positive examples are the instances that belong to the target
class, which in this case are fraudulent transactions. Negative examples are the instances that
belong to the other class, which in this case are legitimate transactions. Fraudulent transactions
are usually rare and imbalanced compared to legitimate transactions, which can cause the model
to be biased towards the majority class and fail to learn the characteristics of the minority class.
Adding more positive examples can help the model learn more features and patterns of the
fraudulent transactions, and increase the detection rate2.
The other options are not as good as options B and C, for the following reasons:
* Option A: Increasing the score threshold would decrease the detection rate of fraudulent
transactions, which is the opposite of the desired outcome. Increasing the score threshold would
decrease the recall of the model, which is the proportion of actual positive cases that are correctly
identified. Decreasing the recall would increase the number of false negatives, which are
fraudulent transactions that are missed by the model. Increasing the score threshold would
increase the precision of the model, which is the proportion of positive predictions that are
actually correct. Increasing the precision would decrease the number of false positives, which are
legitimate transactions that are flagged as fraudulent by the model. However, in this use case, the
cost of false negatives is much higher than the cost of false positives, so increasing the score
threshold is not a good option1.
* Option D: Adding more negative examples to the training set would not improve the model
performance, and could worsen the data imbalance. Negative examples are the instances that
belong to the other class, which in this case are legitimate transactions. Legitimate transactions
are usually abundant and dominant compared to fraudulent transactions, which can cause the
model to be biased towards the majority class and fail to learn the characteristics of the minority
class. Adding more negative examples would exacerbate this problem, and decrease the
detection rate of the fraudulent transactions2.
* Option E: Reducing the maximum number of node hours for training would not improve the
model performance, and could limit the model optimization. Node hours are the units of
computation that are used to train an AutoML model. The maximum number of node hours is a
parameter that determines the upper limit of node hours that can be used for training. Reducing
the maximum number of node hours would reduce the training time and cost, but also the model
quality and accuracy. Reducing the maximum number of node hours would limit the number of
iterations, trials, and evaluations that the model can perform, and prevent the model from finding
the optimal hyperparameters and architecture3.
References:
* Preparing for Google Cloud Certification: Machine Learning Engineer, Course 5: Responsible
AI, Week
4: Evaluation
* Google Cloud Professional Machine Learning Engineer Exam Guide, Section 2: Developing
high-quality ML models, 2.2 Handling imbalanced data
* Official Google Cloud Certified Professional Machine Learning Engineer Study Guide, Chapter
4:
* Low-code ML Solutions, Section 4.3: AutoML
* Understanding the score threshold slider
* Handling imbalanced data sets in machine learning
* AutoML Vision pricing

NEW QUESTION: 13
Your team has been tasked with creating an ML solution in Google Cloud to classify support
requests for one of your platforms. You analyzed the requirements and decided to use
TensorFlow to build the classifier so that you have full control of the model's code, serving, and
deployment. You will use Kubeflow pipelines for the ML platform. To save time, you want to build
on existing resources and use managed services instead of building a completely new model.
How should you build the classifier?
A. Use the Natural Language API to classify support requests
B. Use AutoML Natural Language to build the support requests classifier
C. Use an established text classification model on Al Platform to perform transfer learning
D. Use an established text classification model on Al Platform as-is to classify support requests
Answer: (SHOW ANSWER)
Transfer learning is a technique that leverages the knowledge and weights of a pre-trained model
and adapts them to a new task or domain1. Transfer learning can save time and resources by
avoiding training a model from scratch, and can also improve the performance and generalization
of the model by using a larger and more diverse dataset2. AI Platform provides several
established text classification models that can be used for transfer learning, such as BERT,
ALBERT, or XLNet3. These models are based on state-of-the-art natural language processing
techniques and can handle various text classification tasks, such as sentiment analysis, topic
classification, or spam detection4. By using one of these models on AI Platform, you can
customize the model's code, serving, and deployment, and use Kubeflow pipelines for the ML
platform. Therefore, using an established text classification model on AI Platform to perform
transfer learning is the best option for this use case.
References:
* Transfer Learning - Machine Learning's Next Frontier
* A Comprehensive Hands-on Guide to Transfer Learning with Real-World Applications in Deep
Learning
* Text classification models
* Text Classification with Pre-trained Models in TensorFlow

NEW QUESTION: 14
Your data science team is training a PyTorch model for image classification based on a pre-
trained RestNet model. You need to perform hyperparameter tuning to optimize for several
parameters. What should you do?
A. Convert the model to a Keras model, and run a Keras Tuner job.
B. Run a hyperparameter tuning job on AI Platform using custom containers.
C. Create a Kuberflow Pipelines instance, and run a hyperparameter tuning job on Katib.
D. Convert the model to a TensorFlow model, and run a hyperparameter tuning job on AI
Platform.
Answer: (SHOW ANSWER)
AI Platform supports hyperparameter tuning for PyTorch models using custom containers. This
allows you to use any Python dependencies and libraries that are not included in the pre-built AI
Platform Training runtime versions. You can also use a pre-trained model such as ResNet as a
base for your custom model. To run a hyperparameter tuning job on AI Platform using custom
containers, you need to do the following steps:
* Create a Dockerfile that defines the container image for your training application. The Dockerfile
should install PyTorch and any other dependencies, copy your training code and configuration
files, and set the entrypoint for the container.
* Build the container image and push it to Container Registry or another accessible registry.
* Create a YAML file that defines the configuration for your hyperparameter tuning job. The YAML
file should specify the container image URI, the training input and output paths, the
hyperparameters to tune, the metric to optimize, and the tuning algorithm and budget.
* Submit the hyperparameter tuning job to AI Platform using the gcloud command-line tool or the
AI Platform Training API.
References:
* Hyperparameter tuning overview
* Using custom containers
* PyTorch on AI Platform Training

NEW QUESTION: 15
You work for a retail company that is using a regression model built with BigQuery ML to predict
product sales. This model is being used to serve online predictions Recently you developed a
new version of the model that uses a different architecture (custom model) Initial analysis
revealed that both models are performing as expected You want to deploy the new version of the
model to production and monitor the performance over the next two months You need to minimize
the impact to the existing and future model users How should you deploy the model?
A. Import the new model to the same Vertex Al Model Registry as a different version of the
existing model. Deploy the new model to the same Vertex Al endpoint as the existing model, and
use traffic splitting to route 95% of production traffic to the BigQuery ML model and 5% of
production traffic to the new model.
B. Import the new model to the same Vertex Al Model Registry as the existing model Deploy the
models to one Vertex Al endpoint Route 95% of production traffic to the BigQuery ML model and
5% of production traffic to the new model
C. Import the new model to the same Vertex Al Model Registry as the existing model Deploy each
model to a separate Vertex Al endpoint.
D. Deploy the new model to a separate Vertex Al endpoint Create a Cloud Run service that
routes the prediction requests to the corresponding endpoints based on the input feature values.
Answer: (SHOW ANSWER)
Vertex AI Model Registry is a central repository where you can manage the lifecycle of your ML
models1. You can import models from various sources, such as BigQuery ML, AutoML, or custom
models, and assign them to different versions and aliases1. You can also deploy models to
endpoints, which are resources that provide a service URL for online prediction2.
By importing the new model to the same Vertex AI Model Registry as a different version of the
existing model, you can keep track of the model versions and compare their performance
metrics1. You can also use aliases to label the model versions according to their readiness for
production, such as default or staging1.
By deploying the new model to the same Vertex AI endpoint as the existing model, you can use
traffic splitting to gradually shift the production traffic from the old model to the new model2.
Traffic splitting is a feature that allows you to specify the percentage of prediction requests that
each deployed model in an endpoint should handle2. This way, you can minimize the impact to
the existing and future model users, and monitor the performance of the new model over time2.
The other options are not suitable for your scenario, because they either require creating a
separate endpoint or a Cloud Run service, which would increase the complexity and maintenance
of your deployment, or they do not allow you to use traffic splitting, which would create a sudden
change in your prediction results.
References:
* Introduction to Vertex AI Model Registry | Google Cloud
* Deploy a model to an endpoint | Vertex AI | Google Cloud

NEW QUESTION: 16
You received a training-serving skew alert from a Vertex Al Model Monitoring job running in
production.
You retrained the model with more recent training data, and deployed it back to the Vertex Al
endpoint but you are still receiving the same alert. What should you do?
A. Update the model monitoring job to use a lower sampling rate.
B. Update the model monitoring job to use the more recent training data that was used to retrain
the model.
C. Temporarily disable the alert Enable the alert again after a sufficient amount of new production
traffic has passed through the Vertex Al endpoint.
D. Temporarily disable the alert until the model can be retrained again on newer training data
Retrain the model again after a sufficient amount of new production traffic has passed through the
Vertex Al endpoint
Answer: (SHOW ANSWER)
The best option for resolving the training-serving skew alert is to update the model monitoring job
to use the more recent training data that was used to retrain the model. This option can help align
the baseline distribution of the model monitoring job with the current distribution of the production
data, and eliminate the false positive alerts. Model Monitoring is a service that can track and
compare the results of multiple machine learning runs. Model Monitoring can monitor the model's
prediction input data for feature skew and drift.
Training-serving skew occurs when the feature data distribution in production deviates from the
feature data distribution used to train the model. If the original training data is available, you can
enable skew detection to monitor your models for training-serving skew. Model Monitoring uses
TensorFlow Data Validation (TFDV) to calculate the distributions and distance scores for each
feature, and compares them with a baseline distribution. The baseline distribution is the statistical
distribution of the feature's values in the training data. If the distance score for a feature exceeds
an alerting threshold that you set, Model Monitoring sends you an email alert. However, if you
retrain the model with more recent training data, and deploy it back to the Vertex AI endpoint, the
baseline distribution of the model monitoring job may become outdated and inconsistent with the
current distribution of the production data. This can cause the model monitoring job to generate
false positive alerts, even if the model performance is not deteriorated. To avoid this problem, you
need to update the model monitoring job to use the more recent training data that was used to
retrain the model. This can help the model monitoring job to recalculate the baseline distribution
and the distance scores, and compare them with the current distribution of the production data.
This can also help the model monitoring job to detect any true positive alerts, such as a sudden
change in the production data that causes the model performance to degrade1.
The other options are not as good as option B, for the following reasons:
* Option A: Updating the model monitoring job to use a lower sampling rate would not resolve the
training-serving skew alert, and could reduce the accuracy and reliability of the model monitoring
job.
The sampling rate is a parameter that determines the percentage of prediction requests that are
logged and analyzed by the model monitoring job. Using a lower sampling rate can reduce the
storage and computation costs of the model monitoring job, but also the quality and validity of the
data. Using a lower sampling rate can introduce sampling bias and noise into the data, and make
the model monitoring job miss some important features or patterns of the data. Moreover, using a
lower sampling rate would not address the root cause of the training-serving skew alert, which is
the mismatch between the baseline distribution and the current distribution of the production
data2.
* Option C: Temporarily disabling the alert, and enabling the alert again after a sufficient amount
of new production traffic has passed through the Vertex AI endpoint, would not resolve the
training-serving skew alert, and could expose the model to potential risks and errors. Disabling
the alert would stop the model monitoring job from sending email notifications when the distance
score for a feature exceeds the alerting threshold, but it would not stop the model monitoring job
from calculating and comparing the distributions and distance scores. Therefore, disabling the
alert would not address the root cause of the training-serving skew alert, which is the mismatch
between the baseline distribution and the current distribution of the production data. Moreover,
disabling the alert would prevent the model monitoring job from detecting any true positive alerts,
such as a sudden change in the production data that causes the model performance to degrade.
This can expose the model to potential risks and errors, and affect the user satisfaction and
trust1.
* Option D: Temporarily disabling the alert until the model can be retrained again on newer
training data, and retraining the model again after a sufficient amount of new production traffic
has passed through the Vertex AI endpoint, would not resolve the training-serving skew alert, and
could cause unnecessary costs and efforts. Disabling the alert would stop the model monitoring
job from sending email notifications when the distance score for a feature exceeds the alerting
threshold, but it would not stop the model monitoring job from calculating and comparing the
distributions and distance scores.
Therefore, disabling the alert would not address the root cause of the training-serving skew alert,
which is the mismatch between the baseline distribution and the current distribution of the
production data.
Moreover, disabling the alert would prevent the model monitoring job from detecting any true
positive alerts, such as a sudden change in the production data that causes the model
performance to degrade.
This can expose the model to potential risks and errors, and affect the user satisfaction and trust.
Retraining the model again on newer training data would create a new model version, but it would
not update the model monitoring job to use the newer training data as the baseline distribution.
Therefore, retraining the model again on newer training data would not resolve the training-
serving skew alert, and could cause unnecessary costs and efforts1.
References:
* Preparing for Google Cloud Certification: Machine Learning Engineer, Course 3: Production ML
Systems, Week 4: Evaluation
* Google Cloud Professional Machine Learning Engineer Exam Guide, Section 3: Scaling ML
models in production, 3.3 Monitoring ML models in production
* Official Google Cloud Certified Professional Machine Learning Engineer Study Guide, Chapter
6:
Production ML Systems, Section 6.3: Monitoring ML Models
* Using Model Monitoring
* Understanding the score threshold slider
* Sampling rate

Valid Professional-Machine-Learning-Engineer Dumps shared by ExamDiscuss.com for


Helping Passing Professional-Machine-Learning-Engineer Exam! ExamDiscuss.com now offer
the newest Professional-Machine-Learning-Engineer exam dumps, the ExamDiscuss.com
Professional-Machine-Learning-Engineer exam questions have been updated and answers
have been corrected get the newest ExamDiscuss.com Professional-Machine-Learning-
Engineer dumps with Test Engine here:
https://www.examdiscuss.com/Google/exam/Professional-Machine-Learning-
Engineer/premium/ (273 Q&As Dumps, 35%OFF Special Discount Code: freecram)

NEW QUESTION: 17
You are training and deploying updated versions of a regression model with tabular data by using
Vertex Al Pipelines. Vertex Al Training Vertex Al Experiments and Vertex Al Endpoints. The
model is deployed in a Vertex Al endpoint and your users call the model by using the Vertex Al
endpoint. You want to receive an email when the feature data distribution changes significantly,
so you can retrigger the training pipeline and deploy an updated version of your model What
should you do?
A. Use Vertex Al Model Monitoring Enable prediction drift monitoring on the endpoint. and specify
a notification email.
B. In Cloud Logging, create a logs-based alert using the logs in the Vertex Al endpoint. Configure
Cloud Logging to send an email when the alert is triggered.
C. In Cloud Monitoring create a logs-based metric and a threshold alert for the metric. Configure
Cloud Monitoring to send an email when the alert is triggered.
D. Export the container logs of the endpoint to BigQuery Create a Cloud Function to run a SQL
query over the exported logs and send an email. Use Cloud Scheduler to trigger the Cloud
Function.
Answer: (SHOW ANSWER)
Prediction drift is the change in the distribution of feature values or labels over time. It can affect
the performance and accuracy of the model, and may require retraining or redeploying the model.
Vertex AI Model Monitoring allows you to monitor prediction drift on your deployed models and
endpoints, and set up alerts and notifications when the drift exceeds a certain threshold. You can
specify an email address to receive the notifications, and use the information to retrigger the
training pipeline and deploy an updated version of your model. This is the most direct and
convenient way to achieve your goal. References:
* Vertex AI Model Monitoring
* Monitoring prediction drift
* Setting up alerts and notifications

NEW QUESTION: 18
You need to develop an image classification model by using a large dataset that contains labeled
images in a Cloud Storage Bucket. What should you do?
A. Use Vertex Al Pipelines with the Kubeflow Pipelines SDK to create a pipeline that reads the
images from Cloud Storage and trains the model.
B. Use Vertex Al Pipelines with TensorFlow Extended (TFX) to create a pipeline that reads the
images from Cloud Storage and trams the model.
C. Import the labeled images as a managed dataset in Vertex Al: and use AutoML to tram the
model.
D. Convert the image dataset to a tabular format using Dataflow Load the data into BigQuery and
use BigQuery ML to tram the model.
Answer: (SHOW ANSWER)
The best option for developing an image classification model by using a large dataset that
contains labeled images in a Cloud Storage bucket is to import the labeled images as a managed
dataset in Vertex AI and use AutoML to train the model. This option allows you to leverage the
power and simplicity of Google Cloud to create and deploy a high-quality image classification
model with minimal code and configuration. Vertex AI is a unified platform for building and
deploying machine learning solutions on Google Cloud. Vertex AI can create a managed dataset
from a Cloud Storage bucket that contains labeled images, which can be used to train an AutoML
model. AutoML is a service that can automatically build and optimize machine learning models for
various tasks, such as image classification, object detection, natural language processing, and
tabular data analysis. AutoML can handle the complex aspects of machine learning, such as
feature engineering, model architecture, hyperparameter tuning, and model evaluation. AutoML
can also evaluate, deploy, and monitor the image classification model, and provide online or
batch predictions. By using Vertex AI and AutoML, users can develop an image classification
model by using a large dataset with ease and efficiency.
The other options are not as good as option C, for the following reasons:
* Option A: Using Vertex AI Pipelines with the Kubeflow Pipelines SDK to create a pipeline that
reads the images from Cloud Storage and trains the model would require more skills and steps
than using Vertex AI and AutoML. Vertex AI Pipelines is a service that can orchestrate machine
learning workflows using Vertex AI. Vertex AI Pipelines can run preprocessing and training steps
on custom Docker images, and evaluate, deploy, and monitor the machine learning model.
Kubeflow Pipelines SDK is a Python library that can create and run pipelines on Vertex AI
Pipelines or on Kubeflow, an open-source platform for machine learning on Kubernetes.
However, using Vertex AI Pipelines and Kubeflow Pipelines SDK would require writing code,
building Docker images, defining pipeline components and steps, and managing the pipeline
execution and artifacts. Moreover, Vertex AI Pipelines and Kubeflow Pipelines SDK are not
specialized for image classification, and users would need to use other libraries or frameworks,
such as TensorFlow or PyTorch, to build and train the image classification model.
* Option B: Using Vertex AI Pipelines with TensorFlow Extended (TFX) to create a pipeline that
reads the images from Cloud Storage and trains the model would require more skills and steps
than using Vertex AI and AutoML. TensorFlow Extended (TFX) is a framework that can create
and run end-to-end machine learning pipelines on TensorFlow, a popular library for building and
training deep learning models. TFX can preprocess the data, train and evaluate the model,
validate and push the model, and serve the model for online or batch predictions. However, using
Vertex AI Pipelines and TFX would
* require writing code, building Docker images, defining pipeline components and steps, and
managing the pipeline execution and artifacts. Moreover, TFX is not optimized for image
classification, and users would need to use other libraries or tools, such as TensorFlow Data
Validation, TensorFlow Transform, and TensorFlow Hub, to handle the image data and the model
architecture.
* Option D: Converting the image dataset to a tabular format using Dataflow, loading the data into
BigQuery, and using BigQuery ML to train the model would not handle the image data properly
and could result in a poor model performance. Dataflow is a service that can create scalable and
reliable pipelines to process large volumes of data from various sources. Dataflow can
preprocess the data by using Apache Beam, a programming model for defining and executing
data processing workflows.
BigQuery is a serverless, scalable, and cost-effective data warehouse that can perform fast and
interactive queries on large datasets. BigQuery ML is a service that can create and train machine
learning models by using SQL queries on BigQuery. However, converting the image data to a
tabular format would lose the spatial and semantic information of the images, which are essential
for image classification. Moreover, BigQuery ML is not specialized for image classification, and
users would need to use other tools or techniques, such as feature hashing, embedding, or one-
hot encoding, to handle the categorical features.

NEW QUESTION: 19
You are building an ML model to detect anomalies in real-time sensor data. You will use Pub/Sub
to handle incoming requests. You want to store the results for analytics and visualization. How
should you configure the pipeline?

A. 1 = Dataflow, 2 - Al Platform, 3 = BigQuery


B. 1 = DataProc, 2 = AutoML, 3 = Cloud Bigtable
C. 1 = BigQuery, 2 = AutoML, 3 = Cloud Functions
D. 1 = BigQuery, 2 = Al Platform, 3 = Cloud Storage
Answer: (SHOW ANSWER)
* Dataflow is a fully managed service for executing Apache Beam pipelines that can process
streaming or batch data1.
* Al Platform is a unified platform that enables you to build and run machine learning applications
across Google Cloud2.
* BigQuery is a serverless, highly scalable, and cost-effective cloud data warehouse designed for
business agility3.
These services are suitable for building an ML model to detect anomalies in real-time sensor
data, as they can handle large-scale data ingestion, preprocessing, training, serving, storage, and
visualization. The other options are not as suitable because:
* DataProc is a service for running Apache Spark and Apache Hadoop clusters, which are not
optimized for streaming data processing4.
* AutoML is a suite of machine learning products that enables developers with limited machine
learning expertise to train high-quality models specific to their business needs5. However, it does
not support custom models or real-time predictions.
* Cloud Bigtable is a scalable, fully managed NoSQL database service for large analytical and
operational workloads. However, it is not designed for ad hoc queries or interactive analysis.
* Cloud Functions is a serverless execution environment for building and connecting cloud
services.
However, it is not suitable for storing or visualizing data.
* Cloud Storage is a service for storing and accessing data on Google Cloud. However, it is not a
data warehouse and does not support SQL queries or visualization tools.

NEW QUESTION: 20
A. Use UNIT_LINEAR_SCALE for the embedding dimension, UNIT_LOG_SCALE for the learning
rate, and a large number of parallel trials.
B. Use UNIT_LINEAR_SCALE for the embedding dimension, UNIT_LOG_SCALE for the learning
rate, and a small number of parallel trials.
C. Use UNIT_LOG_SCALE for the embedding dimension, UNIT_LINEAR_SCALE for the learning
rate, and a large number of parallel trials.
D. Use UNIT_LOG_SCALE for the embedding dimension, UNIT_LINEAR_SCALE for the learning
rate, and a small number of parallel trials.
Answer: (SHOW ANSWER)
The best option for performing hyperparameter tuning on Vertex AI to determine the appropriate
embedding dimension and the optimal learning rate is to use UNIT_LINEAR_SCALE for the
embedding dimension, UNIT_LOG_SCALE for the learning rate, and a large number of parallel
trials. This option has the following advantages:
* It matches the appropriate scaling type for each hyperparameter, based on their range and
distribution.
The embedding dimension is an integer hyperparameter that varies linearly between 16 and 64,
so using UNIT_LINEAR_SCALE makes sense. The learning rate is a double hyperparameter that
varies exponentially between 10e-05 and 10e-02, so using UNIT_LOG_SCALE is more suitable.
* It maximizes the exploration of the hyperparameter space, by using a large number of parallel
trials.
Since training time is not a concern, using more trials can help find the best combination of
hyperparameters that maximizes model accuracy. The default Bayesian optimization tuning
algorithm can efficiently sample the hyperparameter space and converge to the optimal values.
The other options are less optimal for the following reasons:
* Option B: Using UNIT_LINEAR_SCALE for the embedding dimension, UNIT_LOG_SCALE for
the learning rate, and a small number of parallel trials, reduces the exploration of the
hyperparameter space, by using a small number of parallel trials. Since training time is not a
concern, using fewer trials can miss some potentially good combinations of hyperparameters that
maximize model accuracy. The default Bayesian optimization tuning algorithm can benefit from
more trials to sample the hyperparameter space and converge to the optimal values.
* Option C: Using UNIT_LOG_SCALE for the embedding dimension, UNIT_LINEAR_SCALE for
the learning rate, and a large number of parallel trials, mismatches the appropriate scaling type
for each hyperparameter, based on their range and distribution. The embedding dimension is an
integer hyperparameter that varies linearly between 16 and 64, so using UNIT_LOG_SCALE is
not suitable.
The learning rate is a double hyperparameter that varies exponentially between 10e-05 and
10e-02, so using UNIT_LINEAR_SCALE makes less sense.
* Option D: Using UNIT_LOG_SCALE for the embedding dimension, UNIT_LINEAR_SCALE for
the learning rate, and a small number of parallel trials, combines the drawbacks of option B and
option C. It mismatches the appropriate scaling type for each hyperparameter, based on their
range and distribution, and reduces the exploration of the hyperparameter space, by using a
small number of parallel trials.
References:
* [Vertex AI: Hyperparameter tuning overview]
* [Vertex AI: Configuring the hyperparameter tuning job]

NEW QUESTION: 21
You are training an LSTM-based model on Al Platform to summarize text using the following job
submission script:

You want to ensure that training time is minimized without significantly compromising the
accuracy of your model. What should you do?
A. Modify the 'epochs' parameter
B. Modify the 'scale-tier' parameter
C. Modify the batch size' parameter
D. Modify the 'learning rate' parameter
Answer: (SHOW ANSWER)
The training time of a machine learning model depends on several factors, such as the complexity
of the model, the size of the data, the hardware resources, and the hyperparameters. To minimize
the training time without significantly compromising the accuracy of the model, one should
optimize these factors as much as possible.
One of the factors that can have a significant impact on the training time is the scale-tier
parameter, which specifies the type and number of machines to use for the training job on AI
Platform. The scale-tier parameter can be one of the predefined values, such as BASIC,
STANDARD_1, PREMIUM_1, or BASIC_GPU, or a custom value that allows you to configure the
machine type, the number of workers, and the number of parameter servers1 To speed up the
training of an LSTM-based model on AI Platform, one should modify the scale-tier parameter to
use a higher tier or a custom configuration that provides more computational resources, such as
more CPUs, GPUs, or TPUs. This can reduce the training time by increasing the parallelism and
throughput of the model training. However, one should also consider the trade-off between the
training time and the cost, as higher tiers or custom configurations may incur higher charges2 The
other options are not as effective or may have adverse effects on the model accuracy. Modifying
the epochs parameter, which specifies the number of times the model sees the entire dataset,
may reduce the training time, but also affect the model's convergence and performance.
Modifying the batch size parameter, which specifies the number of examples per batch, may
affect the model's stability and generalization ability, as well as the memory usage and the
gradient update frequency. Modifying the learning rate parameter, which specifies the step size of
the gradient descent optimization, may affect the model's convergence and performance, as well
as the risk of overshooting or getting stuck in local minima3 References: 1: Using predefined
machine types 2: Distributed training 3: Hyperparameter tuning overview

NEW QUESTION: 22
You are developing models to classify customer support emails. You created models with
TensorFlow Estimators using small datasets on your on-premises system, but you now need to
train the models using large datasets to ensure high performance. You will port your models to
Google Cloud and want to minimize code refactoring and infrastructure overhead for easier
migration from on-prem to cloud. What should you do?
A. Use Vertex Al Platform for distributed training
B. Create a cluster on Dataproc for training
C. Create a Managed Instance Group with autoscaling
D. Use Kubeflow Pipelines to train on a Google Kubernetes Engine cluster.
Answer: (SHOW ANSWER)
Vertex AI Platform is a unified platform for building and deploying ML models on Google Cloud. It
supports both custom and AutoML models, and provides various tools and services for ML
development, such as Vertex Pipelines, Vertex Vizier, Vertex Explainable AI, and Vertex Feature
Store. Vertex AI Platform allows users to train their TensorFlow models using distributed training,
which can speed up the training process and handle large datasets. Vertex AI Platform also
minimizes code refactoring and infrastructure overhead, as it is compatible with TensorFlow
Estimators and handles the provisioning, configuration, and scaling of the training resources
automatically. The other options are not as suitable for this scenario. Dataproc is a service that
allows users to create and run data processing pipelines using Apache Spark and Hadoop, but it
is not designed for TensorFlow model training. Managed Instance Groups are a feature that
allows users to create and manage groups of identical compute instances, but they require more
configuration and management than Vertex AI Platform. Kubeflow Pipelines are a tool that allows
users to create and run ML workflows on Google Kubernetes Engine, but they involve more
complexity and code changes than Vertex AI Platform.
References:
* Vertex AI Platform documentation
* Distributed training with Vertex AI Platform

NEW QUESTION: 23
You have developed a BigQuery ML model that predicts customer churn and deployed the model
to Vertex Al Endpoints. You want to automate the retraining of your model by using minimal
additional code when model feature values change. You also want to minimize the number of
times that your model is retrained to reduce training costs. What should you do?
A. 1. Enable request-response logging on Vertex Al Endpoints.
2 Schedule a TensorFlow Data Validation job to monitor prediction drift
3. Execute model retraining if there is significant distance between the distributions.
B. 1. Enable request-response logging on Vertex Al Endpoints
2. Schedule a TensorFlow Data Validation job to monitor training/serving skew
3. Execute model retraining if there is significant distance between the distributions
C. 1 Create a Vertex Al Model Monitoring job configured to monitor prediction drift.
2. Configure alert monitoring to publish a message to a Pub/Sub queue when a monitonng alert is
detected.
3. Use a Cloud Function to monitor the Pub/Sub queue, and trigger retraining in BigQuery
D. 1. Create a Vertex Al Model Monitoring job configured to monitor training/serving skew
2. Configure alert monitoring to publish a message to a Pub/Sub queue when a monitoring alert is
detected
3. Use a Cloud Function to monitor the Pub/Sub queue, and trigger retraining in BigQuery.
Answer: (SHOW ANSWER)
The best option for automating the retraining of your model by using minimal additional code
when model feature values change, and minimizing the number of times that your model is
retrained to reduce training costs, is to create a Vertex AI Model Monitoring job configured to
monitor prediction drift, configure alert monitoring to publish a message to a Pub/Sub queue when
a monitoring alert is detected, and use a Cloud Function to monitor the Pub/Sub queue, and
trigger retraining in BigQuery. This option allows you to leverage the power and simplicity of
Vertex AI, Pub/Sub, and Cloud Functions to monitor your model performance and retrain your
model when needed. Vertex AI is a unified platform for building and deploying machine learning
solutions on Google Cloud. Vertex AI can deploy a trained model to an online prediction endpoint,
which can provide low-latency predictions for individual instances. Vertex AI can also provide
various tools and services for data analysis, model development, model deployment, model
monitoring, and model governance. A Vertex AI Model Monitoring job is a resource that can
monitor the performance and quality of your deployed models on Vertex AI. A Vertex AI Model
Monitoring job can help you detect and diagnose issues with your models, such as data drift,
prediction drift, training/serving skew, or model staleness. Prediction drift is a type of model
monitoring metric that measures the difference between the distributions of the predictions
generated by the model on the training data and the predictions generated by the model on the
online data. Prediction drift can indicate that the model performance is degrading, or that the
online data is changing over time. By creating a Vertex AI Model Monitoring job configured to
monitor prediction drift, you can track the changes in the model predictions, and compare them
with the expected predictions. Alert monitoring is a feature of Vertex AI Model Monitoring that can
notify you when a monitoring metric exceeds a predefined threshold. Alert monitoring can help
you set up rules and conditions for triggering alerts, and choose the notification channel for
receiving alerts. Pub/Sub is a service that can provide reliable and scalable messaging and event
streaming on Google Cloud. Pub/Sub can help you publish and subscribe to messages, and
deliver them to various Google Cloud services, such as Cloud Functions. A Pub/Sub queue is a
resource that can hold messages that are published to a Pub/Sub topic. A Pub/Sub queue can
help you store and manage messages, and ensure that they are delivered to the subscribers. By
configuring alert monitoring to publish a message to a Pub/Sub queue when a monitoring alert is
detected, you can send a notification to a Pub/Sub topic, and trigger a downstream action based
on the alert. Cloud Functions is a service that can run your stateless code in response to events
on Google Cloud. Cloud Functions can help you create and execute functions without
provisioning or managing servers, and pay only for the resources you use. A Cloud Function is a
resource that can execute a piece of code in response to an event, such as a Pub/Sub message.
A Cloud Function can help you perform various tasks, such as data processing, data
transformation, or data analysis. BigQuery is a service that can store and query large-scale data
on Google Cloud. BigQuery can help you analyze your data by using SQL queries, and perform
various tasks, such as data exploration, data transformation, or data visualization. BigQuery ML is
a feature of BigQuery that can create and execute machine learning models in BigQuery by using
SQL queries.
BigQuery ML can help you build and train various types of models, such as linear regression,
logistic regression, k-means clustering, matrix factorization, and deep neural networks. By using a
Cloud Function to monitor the Pub/Sub queue, and trigger retraining in BigQuery, you can
automate the retraining of your model by using minimal additional code when model feature
values change. You can write a Cloud Function that listens to the Pub/Sub queue, and executes a
SQL query to retrain your model in BigQuery ML when a prediction drift alert is received. By
retraining your model in BigQuery ML, you can update your model parameters and improve your
model performance and accuracy1.
The other options are not as good as option C, for the following reasons:
* Option A: Enabling request-response logging on Vertex AI Endpoints, scheduling a TensorFlow
Data Validation job to monitor prediction drift, and executing model retraining if there is significant
distance between the distributions would require more skills and steps than creating a Vertex AI
Model Monitoring job configured to monitor prediction drift, configuring alert monitoring to publish
a message to a Pub/Sub queue when a monitoring alert is detected, and using a Cloud Function
to monitor the Pub/Sub queue, and trigger retraining in BigQuery. Request-response logging is a
feature of Vertex AI Endpoints that can record the requests and responses that are sent to and
from the online prediction
* endpoint. Request-response logging can help you collect and analyze the online prediction data,
and troubleshoot any issues with your model. TensorFlow Data Validation is a tool that can
analyze and validate your data for machine learning. TensorFlow Data Validation can help you
explore, understand, and clean your data, and detect various data issues, such as data drift, data
skew, or data anomalies.
Prediction drift is a type of data issue that measures the difference between the distributions of
the predictions generated by the model on the training data and the predictions generated by the
model on the online data. Prediction drift can indicate that the model performance is degrading, or
that the online data is changing over time. By enabling request-response logging on Vertex AI
Endpoints, and scheduling a TensorFlow Data Validation job to monitor prediction drift, you can
collect and analyze the online prediction data, and compare the distributions of the predictions.
However, enabling request-response logging on Vertex AI Endpoints, scheduling a TensorFlow
Data Validation job to monitor prediction drift, and executing model retraining if there is significant
distance between the distributions would require more skills and steps than creating a Vertex AI
Model Monitoring job configured to monitor prediction drift, configuring alert monitoring to publish
a message to a Pub/Sub queue when a monitoring alert is detected, and using a Cloud Function
to monitor the Pub/Sub queue, and trigger retraining in BigQuery. You would need to write code,
enable and configure the request-response logging, create and run the TensorFlow Data
Validation job, define and measure the distance between the distributions, and execute the model
retraining. Moreover, this option would not automate the retraining of your model, as you would
need to manually check the prediction drift and trigger the retraining2.
* Option B: Enabling request-response logging on Vertex AI Endpoints, scheduling a TensorFlow
Data Validation job to monitor training/serving skew, and executing model retraining if there is
significant distance between the distributions would not help you monitor the changes in the
model feature values, and could cause errors or poor performance. Training/serving skew is a
type of data issue that measures the difference between the distributions of the features used to
train the model and the features used to serve the model. Training/serving skew can indicate that
the model is not trained on the representative data, or that the data is changing over time. By
enabling request-response logging on Vertex AI Endpoints, and scheduling a TensorFlow Data
Validation job to monitor training/serving skew, you can collect and analyze the online prediction
data, and compare the distributions of the features. However, enabling request-response logging
on Vertex AI Endpoints, scheduling a TensorFlow Data Validation job to monitor training/serving
skew, and executing model retraining if there is significant distance between the distributions
would not help you monitor the changes in the model feature values, and could cause errors or
poor performance. You would need to write code, enable and configure the request-response
logging, create and run the TensorFlow Data Validation job, define and measure the distance
between the distributions, and execute the model retraining. Moreover, this option would not
monitor the prediction drift, which is a more direct and relevant metric for measuring the model
performance and quality2.
* Option D: Creating a Vertex AI Model Monitoring job configured to monitor training/serving
skew, configuring alert monitoring to publish a message to a Pub/Sub queue when a monitoring
alert is detected, and using a Cloud Function to monitor the Pub/Sub queue, and trigger retraining
in BigQuery would not help you monitor the changes in the model feature values, and could
cause errors or poor performance. Training/serving skew is a type of data issue that measures
the difference between the distributions of the features used to train the model and the features
used to serve the model.
Training/serving skew can indicate that the model is not trained on the representative data, or that
the data is changing over time. By creating a Vertex AI Model Monitoring job configured to
monitor training/serving skew, you can track the changes in the model features, and compare
them with the expected features. However, creating a Vertex AI Model Monitoring job configured
to monitor training/serving skew, configuring alert monitoring to publish a message to a Pub/Sub
queue when a monitoring alert is detected, and using a Cloud Function to monitor the Pub/Sub
queue, and trigger
* retraining in BigQuery would not help you monitor the changes in the model feature values, and
could cause errors or poor performance. You would need to write code, create and configure the
Vertex AI Model Monitoring job, configure the alert monitoring, create and configure the Pub/Sub
queue, and write a Cloud Function to trigger the retraining. Moreover, this option would not
monitor the prediction drift, which is a more direct and relevant metric for measuring the model
performance and quality1.
References:
* Preparing for Google Cloud Certification: Machine Learning Engineer, Course 3: Production ML
Systems, Week 4: ML Governance
* Google Cloud Professional Machine Learning Engineer Exam Guide, Section 3: Scaling ML
models in production

NEW QUESTION: 24
You recently used XGBoost to train a model in Python that will be used for online serving Your
model prediction service will be called by a backend service implemented in Golang running on a
Google Kubemetes Engine (GKE) cluster Your model requires pre and postprocessing steps You
need to implement the processing steps so that they run at serving time You want to minimize
code changes and infrastructure maintenance and deploy your model into production as quickly
as possible. What should you do?
A. Use FastAPI to implement an HTTP server Create a Docker image that runs your HTTP server
and deploy it on your organization's GKE cluster.
B. Use FastAPI to implement an HTTP server Create a Docker image that runs your HTTP server
Upload the image to Vertex Al Model Registry and deploy it to a Vertex Al endpoint.
C. Use the Predictor interface to implement a custom prediction routine Build the custom contain
upload the container to Vertex Al Model Registry, and deploy it to a Vertex Al endpoint.
D. Use the XGBoost prebuilt serving container when importing the trained model into Vertex Al
Deploy the model to a Vertex Al endpoint Work with the backend engineers to implement the pre-
and postprocessing steps in the Golang backend service.
Answer: (SHOW ANSWER)
The best option for implementing the processing steps so that they run at serving time, minimizing
code changes and infrastructure maintenance, and deploying the model into production as quickly
as possible, is to use the Predictor interface to implement a custom prediction routine. Build the
custom container, upload the container to Vertex AI Model Registry, and deploy it to a Vertex AI
endpoint. This option allows you to leverage the power and simplicity of Vertex AI to serve your
XGBoost model with minimal effort and customization. Vertex AI is a unified platform for building
and deploying machine learning solutions on Google Cloud. Vertex AI can deploy a trained
XGBoost model to an online prediction endpoint, which can provide low-latency predictions for
individual instances. A custom prediction routine (CPR) is a Python script that defines the logic for
preprocessing the input data, running the prediction, and postprocessing the output data. A CPR
can help you customize the prediction behavior of your model, and handle complex or non-
standard data formats. A CPR can also help you minimize the code changes, as you only need to
write a few functions to implement the prediction logic. A Predictor interface is a class that inherits
from the base class aiplatform.Predictor, and implements the abstract methods predict() and
preprocess(). A Predictor interface can help you create a CPR by defining the preprocessing and
prediction logic for your model. A container image is a package that contains the model, the CPR,
and the dependencies. A container image can help you standardize and simplify the deployment
process, as you only need to upload the container image to Vertex AI Model Registry, and deploy
it to Vertex AI Endpoints. By using the Predictor interface to implement a CPR, building the
custom container, uploading the container to Vertex AI Model Registry, and deploying it to a
Vertex AI endpoint, you can implement the processing steps so that they run at serving time,
minimize code changes and infrastructure maintenance, and deploy the model into production as
quickly as possible1.
The other options are not as good as option C, for the following reasons:
* Option A: Using FastAPI to implement an HTTP server, creating a Docker image that runs your
HTTP server, and deploying it on your organization's GKE cluster would require more skills and
steps than using the Predictor interface to implement a CPR, building the custom container,
uploading the container to Vertex AI Model Registry, and deploying it to a Vertex AI endpoint.
FastAPI is a framework for building web applications and APIs in Python. FastAPI can help you
implement an HTTP server that can handle prediction requests and responses, and perform data
preprocessing and postprocessing. A Docker image is a package that contains the model, the
HTTP server, and the
* dependencies. A Docker image can help you standardize and simplify the deployment process,
as you only need to build and run the Docker image. GKE is a service that can create and
manage Kubernetes clusters on Google Cloud. GKE can help you deploy and scale your Docker
image on Google Cloud, and provide high availability and performance. However, using FastAPI
to implement an HTTP server, creating a Docker image that runs your HTTP server, and
deploying it on your organization's GKE cluster would require more skills and steps than using the
Predictor interface to implement a CPR, building the custom container, uploading the container to
Vertex AI Model Registry, and deploying it to a Vertex AI endpoint. You would need to write code,
create and configure the HTTP server, build and test the Docker image, create and manage the
GKE cluster, and deploy and monitor the Docker image. Moreover, this option would not leverage
the power and simplicity of Vertex AI, which can provide online prediction natively integrated with
Google Cloud services2.
* Option B: Using FastAPI to implement an HTTP server, creating a Docker image that runs your
HTTP server, uploading the image to Vertex AI Model Registry, and deploying it to a Vertex AI
endpoint would require more skills and steps than using the Predictor interface to implement a
CPR, building the custom container, uploading the container to Vertex AI Model Registry, and
deploying it to a Vertex AI endpoint. FastAPI is a framework for building web applications and
APIs in Python. FastAPI can help you implement an HTTP server that can handle prediction
requests and responses, and perform data preprocessing and postprocessing. A Docker image is
a package that contains the model, the HTTP server, and the dependencies. A Docker image can
help you standardize and simplify the deployment process, as you only need to build and run the
Docker image. Vertex AI Model Registry is a service that can store and manage your machine
learning models on Google Cloud. Vertex AI Model Registry can help you upload and organize
your Docker image, and track the model versions and metadata. Vertex AI Endpoints is a service
that can provide online prediction for your machine learning models on Google Cloud. Vertex AI
Endpoints can help you deploy your Docker image to an online prediction endpoint, which can
provide low-latency predictions for individual instances. However, using FastAPI to implement an
HTTP server, creating a Docker image that runs your HTTP server, uploading the image to Vertex
AI Model Registry, and deploying it to a Vertex AI endpoint would require more skills and steps
than using the Predictor interface to implement a CPR, building the custom container, uploading
the container to Vertex AI Model Registry, and deploying it to a Vertex AI endpoint. You would
need to write code, create and configure the HTTP server, build and test the Docker image,
upload the Docker image to Vertex AI Model Registry, and deploy the Docker image to Vertex AI
Endpoints. Moreover, this option would not leverage the power and simplicity of Vertex AI, which
can provide online prediction natively integrated with Google Cloud services2.
* Option D: Using the XGBoost prebuilt serving container when importing the trained model into
Vertex AI, deploying the model to a Vertex AI endpoint, working with the backend engineers to
implement the pre- and postprocessing steps in the Golang backend service would not allow you
to implement the processing steps so that they run at serving time, and could increase the code
changes and infrastructure maintenance. A XGBoost prebuilt serving container is a container
image that is provided by Google Cloud, and contains the XGBoost framework and the
dependencies. A XGBoost prebuilt serving container can help you deploy a XGBoost model
without writing any code, but it also limits your customization options. A XGBoost prebuilt serving
container can only handle standard data formats, such as JSON or CSV, and cannot perform any
preprocessing or postprocessing on the input or output data. If your input data requires any
transformation or normalization before running the prediction, you cannot use a XGBoost prebuilt
serving container. A Golang backend service is a service that is implemented in Golang, a
programming language that can be used for web development and system programming. A
Golang backend service can help you handle the prediction requests and responses from the
frontend, and communicate with the Vertex AI endpoint. However, using the XGBoost prebuilt
serving container when importing the trained model into Vertex AI, deploying the model to a
Vertex AI endpoint, working with the backend engineers to implement the pre- and
postprocessing steps in the
* Golang backend service would not allow you to implement the processing steps so that they run
at serving time, and could increase the code changes and infrastructure maintenance. You would
need to write code, import the trained model into Vertex AI, deploy the model to a Vertex AI
endpoint, implement the pre- and postprocessing steps in the Golang backend service, and test
and monitor the Golang backend service. Moreover, this option would not leverage the power and
simplicity of Vertex AI, which can provide online prediction natively integrated with Google Cloud
services2.
References:
* Preparing for Google Cloud Certification: Machine Learning Engineer, Course 3: Production ML
Systems, Week 2: Serving ML Predictions
* Google Cloud Professional Machine Learning Engineer Exam Guide, Section 3: Scaling ML
models in production, 3.1 Deploying ML models to production
* Official Google Cloud Certified Professional Machine Learning Engineer Study Guide, Chapter
6:
Production ML Systems, Section 6.2: Serving ML Predictions
* Custom prediction routines
* Using pre-built containers for prediction
* Using custom containers for prediction

NEW QUESTION: 25
You work for a social media company. You need to detect whether posted images contain cars.
Each training example is a member of exactly one class. You have trained an object detection
neural network and deployed the model version to Al Platform Prediction for evaluation. Before
deployment, you created an evaluation job and attached it to the Al Platform Prediction model
version. You notice that the precision is lower than your business requirements allow. How should
you adjust the model's final layer softmax threshold to increase precision?
A. Increase the recall
B. Decrease the recall.
C. Increase the number of false positives
D. Decrease the number of false negatives
Answer: (SHOW ANSWER)
Precision and recall are two common metrics for evaluating the performance of a classification
model.
Precision measures the proportion of positive predictions that are correct, while recall measures
the proportion of positive examples that are correctly predicted. Precision and recall are inversely
related, meaning that increasing one will decrease the other, and vice versa. The trade-off
between precision and recall depends on the goal and the cost of the classification problem1.
For the use case of detecting whether posted images contain cars, precision is more important
than recall, as the social media company wants to minimize the number of false positives, or
images that are incorrectly labeled as containing cars. A high precision means that the model is
confident and accurate in its positive predictions, while a low recall means that the model may
miss some positive examples, or images that actually contain cars. The cost of missing some
positive examples is lower than the cost of making wrong positive predictions, as the latter may
affect the user experience and the reputation of the social media company.
The softmax function is a function that transforms a vector of real numbers into a probability
distribution over the possible classes. The softmax function is often used as the final layer of a
neural network for multi-class classification problems, as it assigns a probability to each class,
and the class with the highest probability is chosen as the prediction. The softmax function is
defined as:
softmax (x_i) = exp (x_i) / sum_j exp (x_j)
where x_i is the input value for class i, and softmax (x_i) is the output probability for class i.
The softmax threshold is a parameter that determines the minimum probability that a class must
have to be chosen as the prediction. For example, if the softmax threshold is 0.5, then the class
with the highest probability must have at least 0.5 to be selected, otherwise the prediction is none.
The softmax threshold can be used to adjust the trade-off between precision and recall, as a
higher threshold will increase the precision and decrease the recall, while a lower threshold will
decrease the precision and increase the recall2.
For the use case of detecting whether posted images contain cars, the best way to adjust the
model's final layer softmax threshold to increase precision is to decrease the recall. This means
that the softmax threshold should be increased, so that the model will only make positive
predictions when it is highly confident, and avoid making false positives. By increasing the
softmax threshold, the model will become more selective and accurate in its positive predictions,
and improve the precision metric. Therefore, decreasing the recall is the best option for this use
case.
References:
* Precision and recall - Wikipedia
* How to add a threshold in softmax scores - Stack Overflow

NEW QUESTION: 26
You have deployed multiple versions of an image classification model on Al Platform. You want to
monitor the performance of the model versions overtime. How should you perform this
comparison?
A. Compare the loss performance for each model on a held-out dataset.
B. Compare the loss performance for each model on the validation data
C. Compare the receiver operating characteristic (ROC) curve for each model using the What-lf
Tool
D. Compare the mean average precision across the models using the Continuous Evaluation
feature
Answer: (SHOW ANSWER)
The performance of an image classification model can be measured by various metrics, such as
accuracy, precision, recall, F1-score, and mean average precision (mAP). These metrics can be
calculated based on the confusion matrix, which compares the predicted labels and the true
labels of the images1 One of the best ways to monitor the performance of multiple versions of an
image classification model on AI Platform is to compare the mean average precision across the
models using the Continuous Evaluation feature.
Mean average precision is a metric that summarizes the precision and recall of a model across
different confidence thresholds and classes. Mean average precision is especially useful for multi-
class and multi-label image classification problems, where the model has to assign one or more
labels to each image from a set of possible labels. Mean average precision can range from 0 to 1,
where a higher value indicates a better performance2 Continuous Evaluation is a feature of AI
Platform that allows you to automatically evaluate the performance of your deployed models
using online prediction requests and responses. Continuous Evaluation can help you monitor the
quality and consistency of your models over time, and detect any issues or anomalies that may
affect the model performance. Continuous Evaluation can also provide various evaluation metrics
and visualizations, such as accuracy, precision, recall, F1-score, ROC curve, and confusion
matrix, for different types of models, such as classification, regression, and object detection3 To
compare the mean average precision across the models using the Continuous Evaluation feature,
you need to do the following steps:
* Enable the online prediction logging for each model version that you want to evaluate. This will
allow AI Platform to collect the prediction requests and responses from your models and store
them in BigQuery4
* Create an evaluation job for each model version that you want to evaluate. This will allow AI
Platform to compare the predicted labels and the true labels of the images, and calculate the
evaluation metrics, such as mean average precision. You need to specify the BigQuery table that
contains the prediction logs, the data schema, the label column, and the evaluation interval.
* View the evaluation results for each model version on the AI Platform Models page in the
Google Cloud console. You can see the mean average precision and other metrics for each
model version over time, and compare them using charts and tables. You can also filter the
results by different classes and confidence thresholds.
The other options are not as effective or feasible. Comparing the loss performance for each
model on a held-out dataset or on the validation data is not a good idea, as the loss function may
not reflect the actual performance of the model on the online prediction data, and may vary
depending on the choice of the loss function and the optimization algorithm. Comparing the
receiver operating characteristic (ROC) curve for each model using the What-If Tool is not
possible, as the What-If Tool does not support image data or multi-class classification problems.
References: 1: Confusion matrix 2: Mean average precision 3: Continuous Evaluation overview 4:
Configure online prediction logging : [Create an evaluation job] : [View evaluation results] : [What-
If Tool overview]

NEW QUESTION: 27
You have recently trained a scikit-learn model that you plan to deploy on Vertex Al. This model
will support both online and batch prediction. You need to preprocess input data for model
inference. You want to package the model for deployment while minimizing additional code What
should you do?
A. 1 Upload your model to the Vertex Al Model Registry by using a prebuilt scikit-learn prediction
container
2 Deploy your model to Vertex Al Endpoints, and create a Vertex Al batch prediction job that uses
the instanceConfig.inscanceType setting to transform your input data
B. 1 Wrap your model in a custom prediction routine (CPR). and build a container image from the
CPR local model
2 Upload your sci-kit learn model container to Vertex Al Model Registry
3 Deploy your model to Vertex Al Endpoints, and create a Vertex Al batch prediction job
C. 1. Create a custom container for your sci-kit learn model,
2 Define a custom serving function for your model
3 Upload your model and custom container to Vertex Al Model Registry
4 Deploy your model to Vertex Al Endpoints, and create a Vertex Al batch prediction job
D. 1 Create a custom container for your sci-kit learn model.
2 Upload your model and custom container to Vertex Al Model Registry
3 Deploy your model to Vertex Al Endpoints, and create a Vertex Al batch prediction job that uses
the instanceConfig. instanceType setting to transform your input data
Answer: (SHOW ANSWER)
The best option for deploying a scikit-learn model on Vertex AI with minimal additional code is to
wrap the model in a custom prediction routine (CPR) and build a container image from the CPR
local model. Upload your scikit-learn model container to Vertex AI Model Registry. Deploy your
model to Vertex AI Endpoints, and create a Vertex AI batch prediction job. This option allows you
to leverage the power and simplicity of Google Cloud to deploy and serve a scikit-learn model that
supports both online and batch prediction. Vertex AI is a unified platform for building and
deploying machine learning solutions on Google Cloud. Vertex AI can deploy a trained scikit-learn
model to an online prediction endpoint, which can provide low-latency predictions for individual
instances. Vertex AI can also create a batch prediction job, which can provide high-throughput
predictions for a large batch of instances. A custom prediction routine (CPR) is a Python script
that defines the logic for preprocessing the input data, running the prediction, and postprocessing
the output data. A CPR can help you customize the prediction behavior of your model, and handle
complex or non-standard data formats. A CPR can also help you minimize the additional code, as
you only need to write a few functions to implement the prediction logic. A container image is a
package that contains the model, the CPR, and the dependencies. A container image can help
you standardize and simplify the deployment process, as you only need to upload the container
image to Vertex AI Model Registry, and deploy it to Vertex AI Endpoints. By wrapping the model
in a CPR and building a container image from the CPR local model, uploading the scikit-learn
model container to Vertex AI Model Registry, deploying the model to Vertex AI Endpoints, and
creating a Vertex AI batch prediction job, you can deploy a scikit-learn model on Vertex AI with
minimal additional code1.
The other options are not as good as option B, for the following reasons:
* Option A: Uploading your model to the Vertex AI Model Registry by using a prebuilt scikit-learn
prediction container, deploying your model to Vertex AI Endpoints, and creating a Vertex AI batch
prediction job that uses the instanceConfig.instanceType setting to transform your input data
would not allow you to preprocess the input data for model inference, and could cause errors or
poor performance.
A prebuilt scikit-learn prediction container is a container image that is provided by Google Cloud,
and contains the scikit-learn framework and the dependencies. A prebuilt scikit-learn prediction
container can help you deploy a scikit-learn model without writing any code, but it also limits your
customization options. A prebuilt scikit-learn prediction container can only handle standard data
formats, such as JSON or CSV, and cannot perform any preprocessing or postprocessing on the
input or output data. If your input data requires any transformation or normalization before running
the prediction, you cannot use a prebuilt scikit-learn prediction container. The
instanceConfig.instanceType setting is a parameter that determines the machine type and the
accelerator type for the batch prediction job. The instanceConfig.instanceType setting can help
you optimize the performance and the cost of the batch prediction job, but it cannot help you
transform your input data2.
* Option C: Creating a custom container for your scikit-learn model, defining a custom serving
function for your model, uploading your model and custom container to Vertex AI Model Registry,
and deploying your model to Vertex AI Endpoints, and creating a Vertex AI batch prediction job
would require more skills and steps than using a CPR and a container image. A custom container
is a container image that contains the model, the dependencies, and a web server. A custom
container can help you customize the prediction behavior of your model, and handle complex or
non-standard data formats. A custom serving function is a Python function that defines the logic
for running the prediction on the model. A custom serving function can help you implement the
prediction logic of your model, and handle complex or non-standard data formats. However,
creating a custom container and defining a custom serving function would require more skills and
steps than using a CPR and a container image.
You would need to write code, build and test the container image, configure the web server, and
implement the prediction logic. Moreover, creating a custom container and defining a custom
serving function would not allow you to preprocess the input data for model inference, as the
custom serving function only runs the prediction on the model3.
* Option D: Creating a custom container for your scikit-learn model, uploading your model and
custom container to Vertex AI Model Registry, deploying your model to Vertex AI Endpoints, and
creating a Vertex AI batch prediction job that uses the instanceConfig.instanceType setting to
transform your input
* data would not allow you to preprocess the input data for model inference, and could cause
errors or poor performance. A custom container is a container image that contains the model, the
dependencies, and a web server. A custom container can help you customize the prediction
behavior of your model, and handle complex or non-standard data formats. However, creating a
custom container would require more skills and steps than using a CPR and a container image.
You would need to write code, build and test the container image, and configure the web server.
The instanceConfig.instanceType setting is a parameter that determines the machine type and
the accelerator type for the batch prediction job. The instanceConfig.instanceType setting can
help you optimize the performance and the cost of the batch prediction job, but it cannot help you
transform your input data23.
References:
* Preparing for Google Cloud Certification: Machine Learning Engineer, Course 3: Production ML
Systems, Week 2: Serving ML Predictions
* Google Cloud Professional Machine Learning Engineer Exam Guide, Section 3: Scaling ML
models in production, 3.1 Deploying ML models to production
* Official Google Cloud Certified Professional Machine Learning Engineer Study Guide, Chapter
6:
Production ML Systems, Section 6.2: Serving ML Predictions
* Custom prediction routines
* Using pre-built containers for prediction
* Using custom containers for prediction

NEW QUESTION: 28
You are developing a custom image classification model in Python. You plan to run your training
application on Vertex Al Your input dataset contains several hundred thousand small images You
need to determine how to store and access the images for training. You want to maximize data
throughput and minimize training time while reducing the amount of additional code. What should
you do?
A. Store image files in Cloud Storage and access them directly.
B. Store image files in Cloud Storage and access them by using serialized records.
C. Store image files in Cloud Filestore, and access them by using serialized records.
D. Store image files in Cloud Filestore and access them directly by using an NFS mount point.
Answer: (SHOW ANSWER)
Cloud Storage is a scalable and cost-effective storage service for any type of data. By storing
image files in Cloud Storage, you can access them from anywhere and avoid the overhead of
managing your own storage infrastructure. However, accessing image files directly from Cloud
Storage can be slow and inefficient, especially for large-scale training. A better option is to use
serialized records, such as TFRecord or Apache Avro, which are binary formats that store
multiple images and their labels in a single file. Serialized records can improve the data
throughput and reduce the network latency, as well as enable data compression and sharding.
You can use TensorFlow or Apache Beam APIs to create and read serialized records from Cloud
Storage. This solution requires minimal code changes and can speed up your training time
significantly.
References:
* Cloud Storage | Google Cloud
* TFRecord and tf.Example | TensorFlow Core
* Apache Avro 1.10.2 Specification
* Using Apache Beam with Cloud Storage | Cloud Storage

NEW QUESTION: 29
Your team is building an application for a global bank that will be used by millions of customers.
You built a forecasting model that predicts customers1 account balances 3 days in the future.
Your team will use the results in a new feature that will notify users when their account balance is
likely to drop below $25. How should you serve your predictions?
A. 1. Create a Pub/Sub topic for each user
2 Deploy a Cloud Function that sends a notification when your model predicts that a user's
account balance will drop below the $25 threshold.
B. 1. Create a Pub/Sub topic for each user
2. Deploy an application on the App Engine standard environment that sends a notification when
your model predicts that a user's account balance will drop below the $25 threshold
C. 1. Build a notification system on Firebase
2. Register each user with a user ID on the Firebase Cloud Messaging server, which sends a
notification when the average of all account balance predictions drops below the $25 threshold
D. 1 Build a notification system on Firebase
2. Register each user with a user ID on the Firebase Cloud Messaging server, which sends a
notification when your model predicts that a user's account balance will drop below the $25
threshold
Answer: D (LEAVE A REPLY)
This answer is correct because it uses Firebase, a platform that provides a scalable and reliable
notification system for mobile and web applications. Firebase Cloud Messaging (FCM) allows you
to send messages and notifications to users across different devices and platforms. By registering
each user with a user ID on the FCM server, you can target specific users based on their account
balance predictions and send them personalized notifications when their balance is likely to drop
below the $25 threshold. This way, you can provide a useful and timely feature for your
customers and increase their engagement and retention.
References:
* [Firebase Cloud Messaging]
* [Firebase Cloud Messaging: Send messages to specific devices]

NEW QUESTION: 30
A. Create a Vertex Al hyperparameter tuning job.
B. Train an AutoML image classification model.
C. Create a Vertex Al pipeline that runs different model training jobs in parallel.
D. Create a custom training job that uses the Vertex Al Vizier SDK for parameter optimization.
Answer: (SHOW ANSWER)

NEW QUESTION: 31
You are profiling the performance of your TensorFlow model training time and notice a
performance issue caused by inefficiencies in the input data pipeline for a single 5 terabyte CSV
file dataset on Cloud Storage.
You need to optimize the input pipeline performance. Which action should you try first to increase
the efficiency of your pipeline?
A. Preprocess the input CSV file into a TFRecord file.
B. Randomly select a 10 gigabyte subset of the data to train your model.
C. Split into multiple CSV files and use a parallel interleave transformation.
D. Set the reshuffle_each_iteration parameter to true in the tf.data.Dataset.shuffle method.
Answer: (SHOW ANSWER)
According to the web search results, the TFRecord format is a recommended way to store large
amounts of data efficiently and improve the performance of the data input pipeline123. The
TFRecord format is a binary format that can be compressed and serialized, which reduces the I/O
overhead and the memory footprint of the data1. The tf.data API provides tools to create and read
TFRecord files easily1.
The other options are not as effective as option A. Option B would reduce the amount of data
available for training and might affect the model accuracy. Option C would still require reading
from a single CSV file at a time, which might not utilize the full bandwidth of the remote storage.
Option D would only affect the order of the data elements, not the speed of reading them.

Valid Professional-Machine-Learning-Engineer Dumps shared by ExamDiscuss.com for


Helping Passing Professional-Machine-Learning-Engineer Exam! ExamDiscuss.com now offer
the newest Professional-Machine-Learning-Engineer exam dumps, the ExamDiscuss.com
Professional-Machine-Learning-Engineer exam questions have been updated and answers
have been corrected get the newest ExamDiscuss.com Professional-Machine-Learning-
Engineer dumps with Test Engine here:
https://www.examdiscuss.com/Google/exam/Professional-Machine-Learning-
Engineer/premium/ (273 Q&As Dumps, 35%OFF Special Discount Code: freecram)

NEW QUESTION: 32
You work for a retailer that sells clothes to customers around the world. You have been tasked
with ensuring that ML models are built in a secure manner. Specifically, you need to protect
sensitive customer data that might be used in the models. You have identified four fields
containing sensitive data that are being used by your data science team: AGE,
IS_EXISTING_CUSTOMER, LATITUDE_LONGITUDE, and SHIRT_SIZE.
What should you do with the data before it is made available to the data science team for training
purposes?
A. Tokenize all of the fields using hashed dummy values to replace the real values.
B. Use principal component analysis (PCA) to reduce the four sensitive fields to one PCA vector.
C. Coarsen the data by putting AGE into quantiles and rounding LATITUDE_LONGTTUDE into
single precision. The other two fields are already as coarse as possible.
D. Remove all sensitive data fields, and ask the data science team to build their models using
non-sensitive data.
Answer: (SHOW ANSWER)
The best option for protecting sensitive customer data that might be used in the ML models is to
coarsen the data by putting AGE into quantiles and rounding LATITUDE_LONGITUDE into single
precision. This option has the following advantages:
* It preserves the utility and relevance of the data for the ML models, as the coarsened data still
captures the essential information and patterns that the models need to learn. For example,
putting AGE into quantiles can group the customers into different age ranges, which can be useful
for predicting their preferences or behavior. Rounding LATITUDE_LONGITUDE into single
precision can reduce the precision of the location data, but still retain the general geographic
region of the customers, which can be useful for personalizing the recommendations or offers.
* It reduces the risk of exposing the personal or private information of the customers, as the
coarsened data makes it harder to identify or re-identify the individual customers from the data.
For example, putting AGE into quantiles can hide the exact age of the customers, which can be
considered sensitive or confidential. Rounding LATITUDE_LONGITUDE into single precision can
obscure the exact location of the customers, which can be considered sensitive or confidential.
The other options are less optimal for the following reasons:
* Option A: Tokenizing all of the fields using hashed dummy values to replace the real values
eliminates
* the utility and relevance of the data for the ML models, as the tokenized data loses all the
information and patterns that the models need to learn. For example, tokenizing AGE using
hashed dummy values can make the data meaningless and irrelevant, as the models cannot
learn anything from the random tokens. Tokenizing LATITUDE_LONGITUDE using hashed
dummy values can make the data meaningless and irrelevant, as the models cannot learn
anything from the random tokens.
* Option B: Using principal component analysis (PCA) to reduce the four sensitive fields to one
PCA vector reduces the utility and relevance of the data for the ML models, as the PCA vector
may not capture all the information and patterns that the models need to learn. For example,
using PCA to reduce AGE, IS_EXISTING_CUSTOMER, LATITUDE_LONGITUDE, and
SHIRT_SIZE to one PCA vector can lose some information or introduce noise in the data, as the
PCA vector is a linear combination of the original features, which may not reflect their true
relationship or importance. Moreover, using PCA to reduce the four sensitive fields to one PCA
vector may not reduce the risk of exposing the personal or private information of the customers,
as the PCA vector may still be reversible or linkable to the original data, depending on the amount
of variance explained by the PCA vector and the availability of the PCA transformation matrix.
* Option D: Removing all sensitive data fields, and asking the data science team to build their
models using non-sensitive data reduces the utility and relevance of the data for the ML models,
as the non-sensitive data may not contain enough information and patterns that the models need
to learn. For example, removing AGE, IS_EXISTING_CUSTOMER, LATITUDE_LONGITUDE,
and SHIRT_SIZE from the data can make the data insufficient and unrepresentative, as the
models may not be able to learn the factors that influence the customers' preferences or
behavior. Moreover, removing all sensitive data fields from the data may not be necessary or
feasible, as the data protection legislation may allow the use of sensitive data for the ML models,
as long as the data is processed in a secure and ethical manner, and the customers' consent and
rights are respected.
References:
* Protecting Sensitive Data and AI Models with Confidential Computing | NVIDIA Technical Blog
* Training machine learning models from sensitive data | Fast Data Science
* Securing ML applications. Model security and protection - Medium
* Security of AI/ML systems, ML model security | Cossack Labs
* Vulnerabilities, security and privacy for machine learning models

NEW QUESTION: 33
You work for a large social network service provider whose users post articles and discuss news.
Millions of comments are posted online each day, and more than 200 human moderators
constantly review comments and flag those that are inappropriate. Your team is building an ML
model to help human moderators check content on the platform. The model scores each
comment and flags suspicious comments to be reviewed by a human.
Which metric(s) should you use to monitor the model's performance?
A. Number of messages flagged by the model per minute
B. Number of messages flagged by the model per minute confirmed as being inappropriate by
humans.
C. Precision and recall estimates based on a random sample of 0.1% of raw messages each
minute sent to a human for review
D. Precision and recall estimates based on a sample of messages flagged by the model as
potentially inappropriate each minute
Answer: (SHOW ANSWER)
* Precision measures the fraction of messages flagged by the model that are actually
inappropriate, while recall measures the fraction of inappropriate messages that are flagged by
the model. These metrics are useful for evaluating how well the model can identify and filter out
inappropriate comments.
* Option A is not a good metric because it does not account for the accuracy of the model. The
model might flag many messages that are not inappropriate, or miss many messages that are
inappropriate.
* Option B is better than option A, but it still does not account for the recall of the model. The
model might flag only a few messages that are highly likely to be inappropriate, but miss many
other messages that are less obvious but still inappropriate.
* Option C is not a good metric because it does not focus on the messages that are flagged by
the model.
* The random sample of 0.1% of raw messages might contain very few inappropriate messages,
making the precision and recall estimates unreliable.

NEW QUESTION: 34
You work for an online retailer. Your company has a few thousand short lifecycle products. Your
company has five years of sales data stored in BigQuery. You have been asked to build a model
that will make monthly sales predictions for each product. You want to use a solution that can be
implemented quickly with minimal effort. What should you do?
A. Use Prophet on Vertex Al Training to build a custom model.
B. Use Vertex Al Forecast to build a NN-based model.
C. Use BigQuery ML to build a statistical AR1MA_PLUS model.
D. Use TensorFlow on Vertex Al Training to build a custom model.
Answer: (SHOW ANSWER)
According to the web search results, BigQuery ML1 is a service that allows you to create and
execute machine learning models in BigQuery using SQL queries. BigQuery ML supports various
types of models, such as linear regression, logistic regression, k-means clustering, matrix
factorization, deep neural networks, and time series forecasting1. ARIMA_PLUS2 is a statistical
model for time series forecasting that is built in to BigQuery ML. ARIMA_PLUS stands for
AutoRegressive Integrated Moving Average with eXogenous regressors. ARIMA_PLUS models
the relationship between a target variable and its past values, as well as other external factors
that might influence the target variable. ARIMA_PLUS can handle multiple time series,
seasonality, holidays, and missing values2. Therefore, option C is the best way to use a solution
that can be implemented quickly with minimal effort for the given use case, as it allows you to use
SQL queries to build and run a forecasting model in BigQuery without moving the data or writing
custom code. The other options are not relevant or optimal for this scenario. References:
* BigQuery ML
* ARIMA_PLUS
* Google Professional Machine Learning Certification Exam 2023
* Latest Google Professional Machine Learning Engineer Actual Free Exam Questions

NEW QUESTION: 35
You are developing a custom TensorFlow classification model based on tabular data. Your raw
data is stored in BigQuery contains hundreds of millions of rows, and includes both categorical
and numerical features. You need to use a MaxMin scaler on some numerical features, and apply
a one-hot encoding to some categorical features such as SKU names. Your model will be trained
over multiple epochs. You want to minimize the effort and cost of your solution. What should you
do?
A. 1 Write a SQL query to create a separate lookup table to scale the numerical features.
2. Deploy a TensorFlow-based model from Hugging Face to BigQuery to encode the text
features.
3. Feed the resulting BigQuery view into Vertex Al Training.
B. 1 Use BigQuery to scale the numerical features.
2. Feed the features into Vertex Al Training.
3 Allow TensorFlow to perform the one-hot text encoding.
C. 1 Use TFX components with Dataflow to encode the text features and scale the numerical
features.
2 Export results to Cloud Storage as TFRecords.
3 Feed the data into Vertex Al Training.
D. 1 Write a SQL query to create a separate lookup table to scale the numerical features.
2 Perform the one-hot text encoding in BigQuery.
3. Feed the resulting BigQuery view into Vertex Al Training.
Answer: (SHOW ANSWER)
TFX (TensorFlow Extended) is a platform for end-to-end machine learning pipelines. It provides
components for data ingestion, preprocessing, validation, model training, serving, and monitoring.
Dataflow is a fully managed service for scalable data processing. By using TFX components with
Dataflow, you can perform feature engineering on large-scale tabular data in a distributed and
efficient way. You can use the Transform component to apply the MaxMin scaler and the one-hot
encoding to the numerical and categorical features, respectively. You can also use the
ExampleGen component to read data from BigQuery and the Trainer component to train your
TensorFlow model. The output of the Transform component is a TFRecord file, which is a binary
format for storing TensorFlow data. You can export the TFRecord file to Cloud Storage and feed it
into Vertex AI Training, which is a managed service for training custom machine learning models
on Google Cloud. References:
* TFX | TensorFlow
* Dataflow | Google Cloud
* Vertex AI Training | Google Cloud

NEW QUESTION: 36
You recently designed and built a custom neural network that uses critical dependencies specific
to your organization's framework. You need to train the model using a managed training service
on Google Cloud.
However, the ML framework and related dependencies are not supported by Al Platform Training.
Also, both your model and your data are too large to fit in memory on a single machine. Your ML
framework of choice uses the scheduler, workers, and servers distribution structure. What should
you do?
A. Use a built-in model available on Al Platform Training
B. Build your custom container to run jobs on Al Platform Training
C. Build your custom containers to run distributed training jobs on Al Platform Training
D. Reconfigure your code to a ML framework with dependencies that are supported by Al
Platform Training
Answer: (SHOW ANSWER)
AI Platform Training is a service that allows you to run your machine learning training jobs on
Google Cloud using various features, model architectures, and hyperparameters. You can use AI
Platform Training to scale up your training jobs, leverage distributed training, and access
specialized hardware such as GPUs and TPUs1. AI Platform Training supports several pre-built
containers that provide different ML frameworks and dependencies, such as TensorFlow,
PyTorch, scikit-learn, and XGBoost2. However, if the ML framework and related dependencies
that you need are not supported by the pre-built containers, you can build your own custom
containers and use them to run your training jobs on AI Platform Training3.
Custom containers are Docker images that you create to run your training application. By using
custom containers, you can specify and pre-install all the dependencies needed for your
application, and have full control over the code, serving, and deployment of your model4. Custom
containers also enable you to run distributed training jobs on AI Platform Training, which can help
you train large-scale and complex models faster and more efficiently5. Distributed training is a
technique that splits the training data and computation across multiple machines, and coordinates
them to update the model parameters. AI Platform Training supports two types of distributed
training: parameter server and collective all-reduce. The parameter server architecture consists of
a set of workers that perform the computation, and a set of servers that store and update the
model parameters. The collective all-reduce architecture consists of a set of workers that perform
the computation and synchronize the model parameters among themselves. Both architectures
also have a scheduler that coordinates the workers and servers.
For the use case of training a custom neural network that uses critical dependencies specific to
your organization's framework, the best option is to build your custom containers to run
distributed training jobs on AI Platform Training. This option allows you to use the ML framework
and dependencies of your choice, and train your model on multiple machines without having to
manage the infrastructure. Since your ML framework of choice uses the scheduler, workers, and
servers distribution structure, you can use the parameter server architecture to run your
distributed training job on AI Platform Training. You can specify the number and type of
machines, the custom container image, and the training application arguments when you submit
your training job. Therefore, building your custom containers to run distributed training jobs on AI
Platform Training is the best option for this use case.
References:
* AI Platform Training documentation
* Pre-built containers for training
* Custom containers for training
* Custom containers overview | Vertex AI | Google Cloud
* Distributed training overview
* [Types of distributed training]
* [Distributed training architectures]
* [Using custom containers for training with the parameter server architecture]

NEW QUESTION: 37
You work for an advertising company and want to understand the effectiveness of your
company's latest advertising campaign. You have streamed 500 MB of campaign data into
BigQuery. You want to query the table, and then manipulate the results of that query with a
pandas dataframe in an Al Platform notebook. What should you do?
A. Use Al Platform Notebooks' BigQuery cell magic to query the data, and ingest the results as a
pandas dataframe
B. Export your table as a CSV file from BigQuery to Google Drive, and use the Google Drive API
to ingest the file into your notebook instance
C. Download your table from BigQuery as a local CSV file, and upload it to your Al Platform
notebook instance Use pandas. read_csv to ingest the file as a pandas dataframe
D. From a bash cell in your Al Platform notebook, use the bq extract command to export the table
as a CSV file to Cloud Storage, and then use gsutii cp to copy the data into the notebook Use
pandas. read_csv to ingest the file as a pandas dataframe
Answer: A (LEAVE A REPLY)
AI Platform Notebooks is a service that provides managed Jupyter notebooks for data science
and machine learning. You can use AI Platform Notebooks to create, run, and share your code
and analysis in a collaborative and interactive environment1. BigQuery is a service that allows
you to analyze large-scale and complex data using SQL queries. You can use BigQuery to
stream, store, and query your data in a fast and cost-effective way2. Pandas is a popular Python
library that provides data structures and tools for data analysis and manipulation. You can use
pandas to create, manipulate, and visualize dataframes, which are tabular data structures with
rows and columns3.
AI Platform Notebooks provides a cell magic, %%bigquery, that allows you to run SQL queries on
BigQuery data and ingest the results as a pandas dataframe. A cell magic is a special command
that applies to the whole cell in a Jupyter notebook. The %%bigquery cell magic can take various
arguments, such as the name of the destination dataframe, the name of the destination table in
BigQuery, the project ID, and the query parameters4. By using the %%bigquery cell magic, you
can query the data in BigQuery with minimal code and manipulate the results with pandas in AI
Platform Notebooks. This is the most convenient and efficient way to achieve your goal.
The other options are not as good as option A, because they involve more steps, more code, and
more manual effort. Option B requires you to export your table as a CSV file from BigQuery to
Google Drive, and then use the Google Drive API to ingest the file into your notebook instance.
This option is cumbersome and time-consuming, as it involves moving the data across different
services and formats. Option C requires you to download your table from BigQuery as a local
CSV file, and then upload it to your AI Platform notebook instance. This option is also inefficient
and impractical, as it involves downloading and uploading large files, which can take a long time
and consume a lot of bandwidth. Option D requires you to use a bash cell in your AI Platform
notebook to export the table as a CSV file to Cloud Storage, and then copy the data into the
notebook. This option is also complex and unnecessary, as it involves using different commands
and tools to move the data around. Therefore, option A is the best option for this use case.
References:
* AI Platform Notebooks documentation
* BigQuery documentation
* pandas documentation
* Using Jupyter magics to query BigQuery data

NEW QUESTION: 38
You work as an ML engineer at a social media company, and you are developing a visual filter for
users' profile photos. This requires you to train an ML model to detect bounding boxes around
human faces. You want to use this filter in your company's iOS-based mobile phone application.
You want to minimize code development and want the model to be optimized for inference on
mobile phones. What should you do?
A. Train a model using AutoML Vision and use the "export for Core ML" option.
B. Train a model using AutoML Vision and use the "export for Coral" option.
C. Train a model using AutoML Vision and use the "export for TensorFlow.js" option.
D. Train a custom TensorFlow model and convert it to TensorFlow Lite (TFLite).
Answer: (SHOW ANSWER)
AutoML Vision is a Google Cloud service that allows you to train custom ML models for image
classification, object detection, and segmentation without writing any code. You can use AutoML
Vision to upload your training data, label it, and train a model using a graphical user interface.
You can also evaluate the model's performance and export it for deployment. One of the export
options is Core ML, which is a framework that lets you integrate ML models into iOS applications.
Core ML optimizes the model for on-device performance, power efficiency, and minimal memory
footprint. By using AutoML Vision and Core ML, you can minimize code development and have a
model that is optimized for inference on mobile phones.
References:
* AutoML Vision documentation
* Core ML documentation
NEW QUESTION: 39
You need to quickly build and train a model to predict the sentiment of customer reviews with
custom categories without writing code. You do not have enough data to train a model from
scratch. The resulting model should have high predictive performance. Which service should you
use?
A. AutoML Natural Language
B. Cloud Natural Language API
C. AI Hub pre-made Jupyter Notebooks
D. AI Platform Training built-in algorithms
Answer: (SHOW ANSWER)
AutoML Natural Language is a service that allows you to build and train custom natural language
models without writing code. You can use AutoML Natural Language to perform sentiment
analysis with custom categories, such as positive, negative, or neutral. You can also use pre-
trained models or transfer learning to leverage existing knowledge and reduce the amount of data
required to train a model from scratch. AutoML Natural Language provides a user-friendly
interface and a powerful AutoML engine that optimizes your model for high predictive
performance.
Cloud Natural Language API is a service that provides pre-trained models for common natural
language tasks, such as sentiment analysis, entity analysis, and syntax analysis. However, it
does not allow you to customize the categories or use your own data for training.
AI Hub pre-made Jupyter Notebooks are interactive documents that contain code, text, and
visualizations for various machine learning scenarios. However, they require some coding skills
and data preparation to use them effectively.
AI Platform Training built-in algorithms are pre-configured machine learning algorithms that you
can use to train models on AI Platform. However, they do not support sentiment analysis as a
natural language task.
References:
* AutoML Natural Language documentation
* Cloud Natural Language API documentation
* AI Hub documentation
* AI Platform Training documentation

NEW QUESTION: 40

A.
B.

C.

D.

Answer: C (LEAVE A REPLY)


Vertex Explainable AI is a set of tools and frameworks to help you understand and interpret
predictions made by your machine learning models, natively integrated with a number of Google's
products and services1. With Vertex Explainable AI, you can generate feature-based
explanations that show how much each input feature contributed to the model's prediction2. This
can help you debug and improve your model performance, and build confidence in your model's
behavior. Feature-based explanations are supported for custom image classification models
deployed on Vertex AI Prediction3. References:
* Explainable AI | Google Cloud
* Introduction to Vertex Explainable AI | Vertex AI | Google Cloud
* Supported model types for feature-based explanations | Vertex AI | Google Cloud

NEW QUESTION: 41
You manage a team of data scientists who use a cloud-based backend system to submit training
jobs. This system has become very difficult to administer, and you want to use a managed service
instead. The data scientists you work with use many different frameworks, including Keras,
PyTorch, theano. Scikit-team, and custom libraries. What should you do?
A. Use the Al Platform custom containers feature to receive training jobs using any framework
B. Configure Kubeflow to run on Google Kubernetes Engine and receive training jobs through
TFJob
C. Create a library of VM images on Compute Engine; and publish these images on a centralized
repository
D. Set up Slurm workload manager to receive jobs that can be scheduled to run on your cloud
infrastructure.
Answer: (SHOW ANSWER)
A cloud-based backend system is a system that runs on a cloud platform and provides services or
resources to other applications or users. A cloud-based backend system can be used to submit
training jobs, which are tasks that involve training a machine learning model on a given dataset
using a specific framework and configuration1 However, a cloud-based backend system can also
have some drawbacks, such as:
* High maintenance: A cloud-based backend system may require a lot of administration and
management, such as provisioning, scaling, monitoring, and troubleshooting the cloud resources
and services. This can be time-consuming and costly, and may distract from the core business
objectives2
* Low flexibility: A cloud-based backend system may not support all the frameworks and libraries
that the data scientists need to use for their training jobs. This can limit the choices and
capabilities of the data scientists, and affect the quality and performance of their models3
* Poor integration: A cloud-based backend system may not integrate well with other cloud
services or tools that the data scientists need to use for their machine learning workflows, such as
data processing, model deployment, or model monitoring. This can create compatibility and
interoperability issues, and reduce the efficiency and productivity of the data scientists.
Therefore, it may be better to use a managed service instead of a cloud-based backend system
to submit training jobs. A managed service is a service that is provided and operated by a third-
party provider, and offers various benefits, such as:
* Low maintenance: A managed service handles the administration and management of the cloud
resources and services, and abstracts away the complexity and details of the underlying
infrastructure. This can save time and money, and allow the data scientists to focus on their core
tasks2
* High flexibility: A managed service can support multiple frameworks and libraries that the data
scientists need to use for their training jobs, and allow them to customize and configure their
training environments and parameters. This can enhance the choices and capabilities of the data
scientists, and improve the quality and performance of their models3
* Easy integration: A managed service can integrate seamlessly with other cloud services or tools
that the data scientists need to use for their machine learning workflows, and provide a unified
and consistent interface and experience. This can solve the compatibility and interoperability
issues, and increase the efficiency and productivity of the data scientists.
One of the best options for using a managed service to submit training jobs is to use the AI
Platform custom containers feature to receive training jobs using any framework. AI Platform is a
Google Cloud service that provides a platform for building, deploying, and managing machine
learning models. AI Platform supports various machine learning frameworks, such as
TensorFlow, PyTorch, scikit-learn, and XGBoost, and provides various features, such as
hyperparameter tuning, distributed training, online prediction, and model monitoring.
The AI Platform custom containers feature allows the data scientists to use any framework or
library that they want for their training jobs, and package their training application and
dependencies as a Docker container image. The data scientists can then submit their training
jobs to AI Platform, and specify the container image and the training parameters. AI Platform will
run the training jobs on the cloud infrastructure, and handle the scaling, logging, and monitoring of
the training jobs. The data scientists can also use the AI Platform features to optimize, deploy,
and manage their models.
The other options are not as suitable or feasible. Configuring Kubeflow to run on Google
Kubernetes Engine and receive training jobs through TFJob is not ideal, as Kubeflow is mainly
designed for TensorFlow-based training jobs, and does not support other frameworks or libraries.
Creating a library of VM images on Compute Engine and publishing these images on a
centralized repository is not optimal, as Compute Engine is a low-level service that requires a lot
of administration and management, and does not provide the features and integrations of AI
Platform. Setting up Slurm workload manager to receive jobs that can be scheduled to run on
your cloud infrastructure is not relevant, as Slurm is a tool for managing and scheduling jobs on a
cluster of nodes, and does not provide a managed service for training jobs.
References: 1: Cloud computing 2: Managed services 3: Machine learning frameworks : [Machine
learning workflow] : [AI Platform overview] : [Custom containers for training]

NEW QUESTION: 42
Your company manages an application that aggregates news articles from many different online
sources and sends them to users. You need to build a recommendation model that will suggest
articles to readers that are similar to the articles they are currently reading. Which approach
should you use?
A. Create a collaborative filtering system that recommends articles to a user based on the user's
past behavior.
B. Encode all articles into vectors using word2vec, and build a model that returns articles based
on vector similarity.
C. Build a logistic regression model for each user that predicts whether an article should be
recommended to a user.
D. Manually label a few hundred articles, and then train an SVM classifier based on the manually
classified articles that categorizes additional articles into their respective categories.
Answer: (SHOW ANSWER)
* Option A is incorrect because creating a collaborative filtering system that recommends articles
to a user based on the user's past behavior is not the best approach to suggest articles that are
similar to the articles they are currently reading. Collaborative filtering is a method of
recommendation that uses the ratings or preferences of other users to predict the preferences of
a target user1. However, this method does not consider the content or features of the articles,
and may not be able to find articles that are similar in terms of topic, style, or sentiment.
* Option B is correct because encoding all articles into vectors using word2vec, and building a
model that returns articles based on vector similarity is a suitable approach to suggest articles
that are similar to the
* articles they are currently reading. Word2vec is a technique that learns low-dimensional and
dense representations of words from a large corpus of text, such that words that are semantically
similar have similar vectors2. By applying word2vec to the articles, we can obtain vector
representations of the articles that capture their meaning and usage. Then, we can use a
similarity measure, such as cosine similarity, to find articles that have similar vectors to the
current article3.
* Option C is incorrect because building a logistic regression model for each user that predicts
whether an article should be recommended to a user is not a feasible approach to suggest
articles that are similar to the articles they are currently reading. Logistic regression is a
supervised learning method that models the probability of a binary outcome (such as recommend
or not) based on some input features (such as user profile or article content)4. However, this
method requires a large amount of labeled data for each user, which may not be available or
scalable. Moreover, this method does not directly measure the similarity between articles, but
rather the likelihood of a user's preference.
* Option D is incorrect because manually labeling a few hundred articles, and then training an
SVM classifier based on the manually classified articles that categorizes additional articles into
their respective categories is not an effective approach to suggest articles that are similar to the
articles they are currently reading. SVM (support vector machine) is a supervised learning method
that finds a hyperplane that separates the data into different classes (such as news categories)
with the maximum margin5. However, this method also requires a large amount of labeled data,
which may be costly and time-consuming to obtain. Moreover, this method does not account for
the fine-grained similarity between articles within the same category, or the cross-category
similarity between articles from different categories.
References:
* Collaborative filtering
* Word2vec
* Cosine similarity
* Logistic regression
* SVM

NEW QUESTION: 43
You work with a data engineering team that has developed a pipeline to clean your dataset and
save it in a Cloud Storage bucket. You have created an ML model and want to use the data to
refresh your model as soon as new data is available. As part of your CI/CD workflow, you want to
automatically run a Kubeflow Pipelines training job on Google Kubernetes Engine (GKE). How
should you architect this workflow?
A. Configure your pipeline with Dataflow, which saves the files in Cloud Storage After the file is
saved, start the training job on a GKE cluster
B. Use App Engine to create a lightweight python client that continuously polls Cloud Storage for
new files As soon as a file arrives, initiate the training job
C. Configure a Cloud Storage trigger to send a message to a Pub/Sub topic when a new file is
available in a storage bucket. Use a Pub/Sub-triggered Cloud Function to start the training job on
a GKE cluster
D. Use Cloud Scheduler to schedule jobs at a regular interval. For the first step of the job. check
the timestamp of objects in your Cloud Storage bucket If there are no new files since the last run,
abort the job.
Answer: (SHOW ANSWER)
This option is the best way to architect the workflow, as it allows you to use event-driven and
serverless components to automate the ML training process. Cloud Storage triggers are a feature
that allows you to send notifications to a Pub/Sub topic when an object is created, deleted, or
updated in a storage bucket. Pub/Sub is a service that allows you to publish and subscribe to
messages on various topics. Pub/Sub-triggered Cloud Functions are a type of Cloud Functions
that are invoked when a message is published to a specific Pub/Sub topic. Cloud Functions are a
serverless platform that allows you to run code in response to events. By using these
components, you can create a workflow that starts the training job on a GKE cluster as soon as a
new file is available in the Cloud Storage bucket, without having to manage any servers or poll for
changes. The other options are not as efficient or scalable as this option. Dataflow is a service
that allows you to create and run data processing pipelines, but it is not designed to trigger ML
training jobs on GKE. App Engine is a service that allows you to build and deploy web
applications, but it is not suitable for polling Cloud Storage for new files, as it may incur
unnecessary costs and latency. Cloud Scheduler is a service that allows you to schedule jobs at
regular intervals, but it is not ideal for triggering ML training jobs based on data availability, as it
may miss some files or run unnecessary jobs. References:
* Cloud Storage triggers documentation
* Pub/Sub documentation
* Pub/Sub-triggered Cloud Functions documentation
* Cloud Functions documentation
* Kubeflow Pipelines documentation

NEW QUESTION: 44
You work for a bank and are building a random forest model for fraud detection. You have a
dataset that includes transactions, of which 1% are identified as fraudulent. Which data
transformation strategy would likely improve the performance of your classifier?
A. Write your data in TFRecords.
B. Z-normalize all the numeric features.
C. Oversample the fraudulent transaction 10 times.
D. Use one-hot encoding on all categorical features.
Answer: C (LEAVE A REPLY)
Oversampling is a technique for dealing with imbalanced datasets, where the majority class
dominates the minority class. It balances the distribution of classes by increasing the number of
samples in the minority class.
Oversampling can improve the performance of a classifier by reducing the bias towards the
majority class and increasing the sensitivity to the minority class.
In this case, the dataset includes transactions, of which 1% are identified as fraudulent. This
means that the fraudulent transactions are the minority class and the non-fraudulent transactions
are the majority class. A random forest model trained on this dataset might have a low recall for
the fraudulent transactions, meaning that it might miss many of them and fail to detect fraud. This
could have a high cost for the bank and its customers.
One way to overcome this problem is to oversample the fraudulent transactions 10 times,
meaning that each fraudulent transaction is duplicated 10 times in the training dataset. This would
increase the proportion of fraudulent transactions from 1% to about 10%, making the dataset
more balanced. This would also make the random forest model more aware of the patterns and
features that distinguish fraudulent transactions from non-fraudulent ones, and thus improve its
accuracy and recall for the minority class.
For more information about oversampling and other techniques for imbalanced data, see the
following references:
* Random Oversampling and Undersampling for Imbalanced Classification
* Exploring Oversampling Techniques for Imbalanced Datasets

NEW QUESTION: 45
You work for a gaming company that develops massively multiplayer online (MMO) games. You
built a TensorFlow model that predicts whether players will make in-app purchases of more than
$10 in the next two weeks. The model's predictions will be used to adapt each user's game
experience. User data is stored in BigQuery. How should you serve your model while optimizing
cost, user experience, and ease of management?
A. Import the model into BigQuery ML. Make predictions using batch reading data from BigQuery,
and push the data to Cloud SQL
B. Deploy the model to Vertex AI Prediction. Make predictions using batch reading data from
Cloud Bigtable, and push the data to Cloud SQL.
C. Embed the model in the mobile application. Make predictions after every in-app purchase
event is published in Pub/Sub, and push the data to Cloud SQL.
D. Embed the model in the streaming Dataflow pipeline. Make predictions after every in-app
purchase event is published in Pub/Sub, and push the data to Cloud SQL.
Answer: (SHOW ANSWER)
The best option to serve the model while optimizing cost, user experience, and ease of
management is to deploy the model to Vertex AI Prediction, which is a managed service that can
scale up or down according to the demand and provide low latency and high availability. Vertex AI
Prediction can also handle TensorFlow models natively, without requiring any additional steps or
conversions. By using batch prediction, the model can process large volumes of data efficiently
and periodically, without affecting the user experience. The data can be read from Cloud Bigtable,
which is a scalable and performant NoSQL database that can store user data in a flexible
schema. The predictions can then be pushed to Cloud SQL, which is a fully managed relational
database that can store the predictions in a structured format and enable easy querying and
analysis. This option also simplifies the management of the model and the data, as it leverages
the existing Google Cloud services and does not require any additional infrastructure or code.
The other options are not optimal for the following reasons:
* A. Importing the model into BigQuery ML is not a good option, as it requires converting the
TensorFlow model into a format that BigQuery ML can understand, which can introduce errors
and reduce the performance. Moreover, BigQuery ML is not designed for serving real-time
predictions, but rather for training and evaluating models using SQL queries. Reading and writing
data from BigQuery and Cloud SQL can also incur additional costs and latency, as they are both
relational databases that require schema definition and data transformation.
* C. Embedding the model in the mobile application is not a good option, as it increases the size
and complexity of the application, and requires updating the application every time the model
changes.
Moreover, it exposes the model to the users, which can pose security and privacy risks, as well
as potential misuse or abuse. Additionally, it does not leverage the benefits of the cloud, such as
scalability, reliability, and performance.
* D. Embedding the model in the streaming Dataflow pipeline is not a good option, as it requires
building and maintaining a custom pipeline that can handle the model inference and data
processing. This can increase the development and operational costs and complexity, as well as
the potential for errors and failures. Moreover, it does not take advantage of the batch prediction
feature of Vertex AI Prediction, which can optimize the resource utilization and cost efficiency.
References:
* Professional ML Engineer Exam Guide
* Preparing for Google Cloud Certification: Machine Learning Engineer Professional Certificate
* Google Cloud launches machine learning engineer certification
* Vertex AI Prediction documentation
* Cloud Bigtable documentation
* Cloud SQL documentation

NEW QUESTION: 46
You work on an operations team at an international company that manages a large fleet of on-
premises servers located in few data centers around the world. Your team collects monitoring
data from the servers, including CPU/memory consumption. When an incident occurs on a server,
your team is responsible for fixing it.
Incident data has not been properly labeled yet. Your management team wants you to build a
predictive maintenance solution that uses monitoring data from the VMs to detect potential
failures and then alerts the service desk team. What should you do first?
A. Train a time-series model to predict the machines' performance values. Configure an alert if a
machine's actual performance values significantly differ from the predicted performance values.
B. Implement a simple heuristic (e.g., based on z-score) to label the machines' historical
performance data.
Train a model to predict anomalies based on this labeled dataset.
C. Develop a simple heuristic (e.g., based on z-score) to label the machines' historical
performance data.
Test this heuristic in a production environment.
D. Hire a team of qualified analysts to review and label the machines' historical performance data.
Train a model based on this manually labeled dataset.
Answer: (SHOW ANSWER)
* Option A is incorrect because training a time-series model to predict the machines' performance
values, and configuring an alert if a machine's actual performance values significantly differ from
the predicted performance values, is not the best way to build a predictive maintenance solution
that uses monitoring data from the VMs to detect potential failures and then alerts the service
desk team. This option assumes that the performance values follow a predictable pattern, which
may not be the case for complex systems. Moreover, this option does not use any historical
incident data, which may contain useful information for identifying failures. Furthermore, this
option does not involve any model evaluation or validation, which are essential steps for ensuring
the quality and reliability of the model.
* Option B is correct because implementing a simple heuristic (e.g., based on z-score) to label the
machines' historical performance data, and training a model to predict anomalies based on this
labeled dataset, is a reasonable way to build a predictive maintenance solution that uses
monitoring data from the VMs to detect potential failures and then alerts the service desk team.
This option uses a simple and fast method to label the historical performance data, which is
necessary for supervised learning. A z-score is a measure of how many standard deviations a
value is away from the mean of a distribution1.
By using a z-score, we can label the performance values that are unusually high or low as
anomalies, which may indicate failures. Then, we can train a model to learn the patterns of
normal and anomalous performance values, and use it to predict anomalies on new data. We can
also evaluate and validate the model using metrics such as precision, recall, or F1-score, and
compare it with other models or methods.
* Option C is incorrect because developing a simple heuristic (e.g., based on z-score) to label the
machines' historical performance data, and testing this heuristic in a production environment, is
not a safe way to build a predictive maintenance solution that uses monitoring data from the VMs
to detect potential failures and then alerts the service desk team. This option does not involve any
model training or evaluation, which are essential steps for ensuring the quality and reliability of
the solution. Moreover, this option does not test the heuristic on a separate dataset, such as a
validation or test set, before deploying it to production, which may lead to errors or failures in the
production environment.
* Option D is incorrect because hiring a team of qualified analysts to review and label the
machines' historical performance data, and training a model based on this manually labeled
dataset, is not a feasible way to build a predictive maintenance solution that uses monitoring data
from the VMs to detect potential failures and then alerts the service desk team. This option may
produce high-quality labels, but it is also costly, time-consuming, and prone to human errors or
biases. Moreover, this option may not scale well with large or complex datasets, which may
require more analysts or more time to label.
References:
* Z-score
* [Predictive maintenance]
* [Anomaly detection]
* [Time-series analysis]
* [Model evaluation]
Valid Professional-Machine-Learning-Engineer Dumps shared by ExamDiscuss.com for
Helping Passing Professional-Machine-Learning-Engineer Exam! ExamDiscuss.com now offer
the newest Professional-Machine-Learning-Engineer exam dumps, the ExamDiscuss.com
Professional-Machine-Learning-Engineer exam questions have been updated and answers
have been corrected get the newest ExamDiscuss.com Professional-Machine-Learning-
Engineer dumps with Test Engine here:
https://www.examdiscuss.com/Google/exam/Professional-Machine-Learning-
Engineer/premium/ (273 Q&As Dumps, 35%OFF Special Discount Code: freecram)

NEW QUESTION: 47
You work at a bank You have a custom tabular ML model that was provided by the bank's
vendor. The training data is not available due to its sensitivity. The model is packaged as a Vertex
Al Model serving container which accepts a string as input for each prediction instance. In each
string the feature values are separated by commas. You want to deploy this model to production
for online predictions, and monitor the feature distribution over time with minimal effort What
should you do?
A. 1 Upload the model to Vertex Al Model Registry and deploy the model to a Vertex Ai endpoint.
2. Create a Vertex Al Model Monitoring job with feature drift detection as the monitoring objective,
and provide an instance schema.
B. 1 Upload the model to Vertex Al Model Registry and deploy the model to a Vertex Al endpoint.
2 Create a Vertex Al Model Monitoring job with feature skew detection as the monitoring objective
and provide an instance schema.
C. 1 Refactor the serving container to accept key-value pairs as input format.
2. Upload the model to Vertex Al Model Registry and deploy the model to a Vertex Al endpoint.
3. Create a Vertex Al Model Monitoring job with feature drift detection as the monitoring objective.
D. 1 Refactor the serving container to accept key-value pairs as input format.
2 Upload the model to Vertex Al Model Registry and deploy the model to a Vertex Al endpoint.
3. Create a Vertex Al Model Monitoring job with feature skew detection as the monitoring
objective.
Answer: (SHOW ANSWER)
The best option for deploying a custom tabular ML model to production for online predictions, and
monitoring the feature distribution over time with minimal effort, using a model that was provided
by the bank's vendor, the training data is not available due to its sensitivity, and the model is
packaged as a Vertex AI Model serving container which accepts a string as input for each
prediction instance, is to upload the model to Vertex AI Model Registry and deploy the model to a
Vertex AI endpoint, create a Vertex AI Model Monitoring job with feature drift detection as the
monitoring objective, and provide an instance schema. This option allows you to leverage the
power and simplicity of Vertex AI to serve and monitor your model with minimal code and
configuration. Vertex AI is a unified platform for building and deploying machine learning solutions
on Google Cloud. Vertex AI can deploy a trained model to an online prediction endpoint, which
can provide low-latency predictions for individual instances. Vertex AI can also provide various
tools and services for data analysis, model development, model deployment, model monitoring,
and model governance. A Vertex AI Model Registry is a resource that can store and manage your
models on Vertex AI. A Vertex AI Model Registry can help you organize and track your models,
and access various model information, such as model name, model description, and model
labels. A Vertex AI Model serving container is a resource that can run your custom model code on
Vertex AI. A Vertex AI Model serving container can help you package your model code and
dependencies into a container image, and deploy the container image to an online prediction
endpoint. A Vertex AI Model serving container can accept various input formats, such as JSON,
CSV, or TFRecord. A string input format is a type of input format that accepts a string as input for
each prediction instance. A string input format can help you encode your feature values into a
single string, and separate them by commas. By uploading the model to Vertex AI Model Registry
and deploying the model to a Vertex AI endpoint, you can serve your model for online predictions
with minimal code and configuration. You can use the Vertex AI API or the gcloud command-line
tool to upload the model to Vertex AI Model Registry, and provide the model name, model
description, and model labels. You can also use the Vertex AI API or the gcloud command-line
tool to deploy the model to a Vertex AI endpoint, and provide the endpoint name, endpoint
description, endpoint labels, and endpoint resources. A Vertex AI Model Monitoring job is a
resource that can monitor the performance and quality of your deployed models on Vertex AI. A
Vertex AI Model Monitoring job can help you detect and diagnose issues with your models, such
as data drift, prediction drift, training/serving skew, or model staleness. Feature drift is a type of
model monitoring metric that measures the difference between the distributions of the features
used to train the model and the features used to serve the model over time. Feature drift can
indicate that the online data is changing over time, and the model performance is degrading. By
creating a Vertex AI Model Monitoring job with feature drift detection as the monitoring objective,
and providing an instance schema, you can monitor the feature distribution over time with minimal
effort. You can use the Vertex AI API or the gcloud command-line tool to create a Vertex AI Model
Monitoring job, and provide the monitoring objective, the monitoring frequency, the alerting
threshold, and the notification channel. You can also provide an instance schema, which is a
JSON file that describes the features and their types in the prediction input data. An instance
schema can help Vertex AI Model Monitoring parse and analyze the string input format, and
calculate the feature distributions and distance scores1.
The other options are not as good as option A, for the following reasons:
* Option B: Uploading the model to Vertex AI Model Registry and deploying the model to a Vertex
AI endpoint, creating a Vertex AI Model Monitoring job with feature skew detection as the
monitoring objective, and providing an instance schema would not help you monitor the changes
in the online data over time, and could cause errors or poor performance. Feature skew is a type
of model monitoring metric that measures the difference between the distributions of the features
used to train the model and the features used to serve the model at a given point in time. Feature
skew can indicate that the model is not trained on the representative data, or that the data is
changing over time. By creating a Vertex AI Model Monitoring job with feature skew detection as
the monitoring objective, and providing an instance schema, you can monitor the feature
distribution at a given point in time with minimal effort.
However, uploading the model to Vertex AI Model Registry and deploying the model to a Vertex
AI endpoint, creating a Vertex AI Model Monitoring job with feature skew detection as the
monitoring objective, and providing an instance schema would not help you monitor the changes
in the online data over time, and could cause errors or poor performance. You would need to use
the Vertex AI API or the gcloud command-line tool to upload the model to Vertex AI Model
Registry, deploy the model to a
* Vertex AI endpoint, create a Vertex AI Model Monitoring job, and provide an instance schema.
Moreover, this option would not monitor the feature drift, which is a more direct and relevant
metric for measuring the changes in the online data over time, and the model performance and
quality1.
* Option C: Refactoring the serving container to accept key-value pairs as input format, uploading
the model to Vertex AI Model Registry and deploying the model to a Vertex AI endpoint, creating
a Vertex AI Model Monitoring job with feature drift detection as the monitoring objective would
require more skills and steps than uploading the model to Vertex AI Model Registry and deploying
the model to a Vertex AI endpoint, creating a Vertex AI Model Monitoring job with feature drift
detection as the monitoring objective, and providing an instance schema. A key-value pair input
format is a type of input format that accepts a key-value pair as input for each prediction instance.
A key-value pair input format can help you specify the feature names and values in a JSON
object, and separate them by colons. By refactoring the serving container to accept key-value
pairs as input format, uploading the model to Vertex AI Model Registry and deploying the model
to a Vertex AI endpoint, creating a Vertex AI Model Monitoring job with feature drift detection as
the monitoring objective, you can serve and monitor your model with minimal code and
configuration. You can write code to refactor the serving container to accept key-value pairs as
input format, and use the Vertex AI API or the gcloud command-line tool to upload the model to
Vertex AI Model Registry, deploy the model to a Vertex AI endpoint, and create a Vertex AI Model
Monitoring job. However, refactoring the serving container to accept key-value pairs as input
format, uploading the model to Vertex AI Model Registry and deploying the model to a Vertex AI
endpoint, creating a Vertex AI Model Monitoring job with feature drift detection as the monitoring
objective would require more skills and steps than uploading the model to Vertex AI Model
Registry and deploying the model to a Vertex AI endpoint, creating a Vertex AI Model Monitoring
job with feature drift detection as the monitoring objective, and providing an instance schema.
You would need to write code, refactor the serving container, upload the model to Vertex AI
Model Registry, deploy the model to a Vertex AI endpoint, and create a Vertex AI Model
Monitoring job. Moreover, this option would not use the instance schema, which is a JSON file
that can help Vertex AI Model Monitoring parse and analyze the string input format, and calculate
the feature distributions and distance scores1.
* Option D: Refactoring the serving container to accept key-value pairs as input format, uploading
the model to Vertex AI Model Registry and deploying the model to a Vertex AI endpoint, creating
a Vertex AI Model Monitoring job with feature skew detection as the monitoring objective would
require more skills and steps than uploading the model to Vertex AI Model Registry and deploying
the model to a Vertex AI endpoint, creating a Vertex AI Model Monitoring job with feature drift
detection as the monitoring objective, and providing an instance schema, and would not help you
monitor the changes in the online data over time, and could cause errors or poor performance.
Feature skew is a type of model monitoring metric that measures the difference between the
distributions of the features used to train the model and the features used to serve the model at a
given point in time. Feature skew can indicate that the model is not trained on the representative
data, or that the data is changing over time. By creating a Vertex AI Model Monitoring job with
feature skew detection as the monitoring objective, you can monitor the feature distribution at a
given point in time with minimal effort. However, refactoring the serving container to accept key-
value pairs as input format, uploading the model to Vertex AI Model Registry and deploying the
model to a Vertex AI endpoint, creating a Vertex AI Model Monitoring job with feature skew
detection as the monitoring objective would require more skills and steps than uploading the
model to Vertex AI Model Registry and deploying the model to a Vertex AI endpoint, creating a
Vertex AI Model Monitoring job with feature drift detection as the monitoring objective, and
providing an instance schema, and would not help you monitor the changes in the online data
over time, and could cause errors or poor performance. You would need to write code, refactor
the serving container, upload the model to Vertex AI Model Registry, deploy the model to a Vertex
AI endpoint, and create a Vertex AI Model Monitoring job. Moreover, this option would not monitor
the feature drift,
* which is a more direct and relevant metric for measuring the changes in the online data over
time, and the model performance and quality1.
References:
* Using Model Monitoring | Vertex AI | Google Cloud

NEW QUESTION: 48
You have a demand forecasting pipeline in production that uses Dataflow to preprocess raw data
prior to model training and prediction. During preprocessing, you employ Z-score normalization on
data stored in BigQuery and write it back to BigQuery. New training data is added every week.
You want to make the process more efficient by minimizing computation time and manual
intervention. What should you do?
A. Normalize the data using Google Kubernetes Engine
B. Translate the normalization algorithm into SQL for use with BigQuery
C. Use the normalizer_fn argument in TensorFlow's Feature Column API
D. Normalize the data with Apache Spark using the Dataproc connector for BigQuery
Answer: (SHOW ANSWER)
Z-score normalization is a technique that transforms the values of a numeric variable into
standardized units, such that the mean is zero and the standard deviation is one. Z-score
normalization can help to compare variables with different scales and ranges, and to reduce the
effect of outliers and skewness. The formula for z-score normalization is:
z = (x - mu) / sigma
where x is the original value, mu is the mean of the variable, and sigma is the standard deviation
of the variable.
Dataflow is a service that allows you to create and run data processing pipelines on Google
Cloud. You can use Dataflow to preprocess raw data prior to model training and prediction, such
as applying z-score normalization on data stored in BigQuery. However, using Dataflow for this
task may not be the most efficient option, as it involves reading and writing data from and to
BigQuery, which can be time-consuming and costly. Moreover, using Dataflow requires manual
intervention to update the pipeline whenever new training data is added.
A more efficient way to perform z-score normalization on data stored in BigQuery is to translate
the normalization algorithm into SQL and use it with BigQuery. BigQuery is a service that allows
you to analyze large-scale and complex data using SQL queries. You can use BigQuery to
perform z-score normalization on your data using SQL functions such as AVG(),
STDDEV_POP(), and OVER(). For example, the following SQL query can normalize the values of
a column called temperature in a table called weather:
SELECT (temperature - AVG(temperature) OVER ()) / STDDEV_POP(temperature) OVER () AS
normalized_temperature FROM weather; By using SQL to perform z-score normalization on
BigQuery, you can make the process more efficient by minimizing computation time and manual
intervention. You can also leverage the scalability and performance of BigQuery to handle large
and complex datasets. Therefore, translating the normalization algorithm into SQL for use with
BigQuery is the best option for this use case.

NEW QUESTION: 49
You recently built the first version of an image segmentation model for a self-driving car. After
deploying the model, you observe a decrease in the area under the curve (AUC) metric. When
analyzing the video recordings, you also discover that the model fails in highly congested traffic
but works as expected when there is less traffic. What is the most likely reason for this result?
A. The model is overfitting in areas with less traffic and underfitting in areas with more traffic.
B. AUC is not the correct metric to evaluate this classification model.
C. Too much data representing congested areas was used for model training.
D. Gradients become small and vanish while backpropagating from the output to input nodes.
Answer: (SHOW ANSWER)
The most likely reason for the observed result is that the model is overfitting in areas with less
traffic and underfitting in areas with more traffic. Overfitting means that the model learns the
specific patterns and noise in the training data, but fails to generalize well to new and unseen
data. Underfitting means that the model is not able to capture the complexity and variability of the
data, and performs poorly on both training and test data. In this case, the model might have
learned to segment the images well when there is less traffic, but it might not have enough data
or features to handle the more challenging scenarios when there is more traffic.
This could lead to a decrease in the AUC metric, which measures the ability of the model to
distinguish between different classes. AUC is a suitable metric for this classification model, as it is
not affected by class imbalance or threshold selection. The other options are not likely to be the
reason for the result, as they are not related to the traffic density. Too much data representing
congested areas would not cause the model to fail in those areas, but rather help the model learn
better. Gradients vanishing or exploding is a problem that occurs during the training process, not
after the deployment, and it affects the whole model, not specific scenarios.
References:
* Image Segmentation: U-Net For Self Driving Cars
* Intelligent Semantic Segmentation for Self-Driving Vehicles Using Deep Learning
* Sharing Pixelopolis, a self-driving car demo from Google I/O built with TensorFlow Lite
* Google Cloud launches machine learning engineer certification
* Google Professional Machine Learning Engineer Certification
* Professional ML Engineer Exam Guide
* Preparing for Google Cloud Certification: Machine Learning Engineer Professional Certificate

NEW QUESTION: 50

A. Write the transformations into Spark that uses the spark-bigquery-connector and use Dataproc
to preprocess the data.
B. Write SQL queries to transform the data in-place in BigQuery.
C. Add the transformations as a preprocessing layer in the TensorFlow models.
D. Create a Dataflow pipeline that uses the BigQuerylO connector to ingest the data process it
and write it back to BigQuery.
Answer: (SHOW ANSWER)
The best option for configuring the workflow is to add the transformations as a preprocessing
layer in the TensorFlow models. This option allows you to leverage the power and simplicity of
TensorFlow to preprocess and transform the data with simple Python code. TensorFlow is a
framework for building and training machine learning models. TensorFlow provides various tools
and libraries for data analysis and machine learning. A preprocessing layer is a type of layer in
TensorFlow that can perform data preprocessing and feature engineering operations on the input
data. A preprocessing layer can help you customize the data transformation and preprocessing
logic, and handle complex or non-standard data formats. A preprocessing layer can also help you
minimize the preprocessing time, cost, and development effort, as you only need to write a few
lines of code to implement the preprocessing layer, and you do not need to create any
intermediate data sources or pipelines. By adding the transformations as a preprocessing layer in
the TensorFlow models, you can use Vertex AI's custom training service to train multiple
TensorFlow models that read the data from BigQuery and predict future sales1.
The other options are not as good as option C, for the following reasons:
* Option A: Writing the transformations into Spark that uses the spark-bigquery-connector and
using Dataproc to preprocess the data would require more skills and steps than using a
preprocessing layer in TensorFlow. Spark is a framework for distributed data processing and
machine learning. Spark can read and write data from BigQuery by using the spark-bigquery-
connector, which is a library that allows Spark to communicate with BigQuery. Dataproc is a
service that can create and manage Spark clusters on Google Cloud. Dataproc can help you run
Spark jobs on Google Cloud, and scale the clusters according to the workload. However, writing
the transformations into Spark that uses the spark-bigquery-connector and using Dataproc to
preprocess the data would require more skills and steps than using a preprocessing layer in
TensorFlow. You would need to write code, create and configure the Spark cluster, install and
import the spark-bigquery-connector, load and preprocess the data, and write the data back to
BigQuery. Moreover, this option would create an intermediate data source in BigQuery, which can
increase the storage and computation costs2.
* Option B: Writing SQL queries to transform the data in-place in BigQuery would not allow you to
use Vertex AI's custom training service to train multiple TensorFlow models that read the data
from BigQuery and predict future sales. BigQuery is a service that can perform data analysis and
machine learning by using SQL queries. BigQuery can perform data transformation and
preprocessing by using SQL functions and clauses, such as MIN, MAX, CASE, and
TRANSFORM. BigQuery can also perform machine learning by using BigQuery ML, which is a
feature that can create and train machine learning models by using SQL queries. However,
writing SQL queries to transform the data in-place in BigQuery would not allow you to use Vertex
AI's custom training service to train multiple TensorFlow models that read the data from BigQuery
and predict future sales. Vertex AI's custom training service is a service that can run your custom
machine learning code on Vertex AI. Vertex AI's custom training service can support various
machine learning frameworks, such as TensorFlow, PyTorch, and scikit-learn. Vertex AI's custom
training service cannot support SQL queries, as SQL is not a machine
* learning framework. Therefore, if you want to use Vertex AI's custom training service, you
cannot use SQL queries to transform the data in-place in BigQuery3.
* Option D: Creating a Dataflow pipeline that uses the BigQueryIO connector to ingest the data,
process it, and write it back to BigQuery would require more skills and steps than using a
preprocessing layer in TensorFlow. Dataflow is a service that can create and run data processing
and machine learning pipelines on Google Cloud. Dataflow can read and write data from
BigQuery by using the BigQueryIO connector, which is a library that allows Dataflow to
communicate with BigQuery. Dataflow can perform data transformation and preprocessing by
using Apache Beam, which is a framework for distributed data processing and machine learning.
However, creating a Dataflow pipeline that uses the BigQueryIO connector to ingest the data,
process it, and write it back to BigQuery would require more skills and steps than using a
preprocessing layer in TensorFlow. You would need to write code, create and configure the
Dataflow pipeline, install and import the BigQueryIO connector, load and preprocess the data,
and write the data back to BigQuery. Moreover, this option would create an intermediate data
source in BigQuery, which can increase the storage and computation costs4.
References:
* Preparing for Google Cloud Certification: Machine Learning Engineer, Course 3: Production ML
Systems, Week 2: Serving ML Predictions
* Google Cloud Professional Machine Learning Engineer Exam Guide, Section 2: Developing ML
models, 2.1 Developing ML models by using TensorFlow
* Official Google Cloud Certified Professional Machine Learning Engineer Study Guide, Chapter
4:
Developing ML Models, Section 4.1: Developing ML Models by Using TensorFlow
* TensorFlow Preprocessing Layers
* Spark and BigQuery
* Dataproc
* BigQuery ML
* Dataflow and BigQuery
* Apache Beam

NEW QUESTION: 51
You work at a gaming startup that has several terabytes of structured data in Cloud Storage. This
data includes gameplay time data, user metadata, and game metadata. You want to build a
model that recommends new games to users that requires the least amount of coding. What
should you do?
A. Load the data in BigQuery. Use BigQuery ML to train an Autoencoder model.
B. Load the data in BigQuery. Use BigQuery ML to train a matrix factorization model.
C. Read data to a Vertex Al Workbench notebook. Use TensorFlow to train a two-tower model.
D. Read data to a Vertex Al Workbench notebook. Use TensorFlow to train a matrix factorization
model.
Answer: (SHOW ANSWER)
The best option to build a game recommendation model with the least amount of coding is to use
BigQuery ML, which allows you to create and execute machine learning models using standard
SQL queries. BigQuery ML supports several types of models, including matrix factorization, which
is a common technique for collaborative filtering-based recommendation systems. Matrix
factorization models learn latent factors for users and items from the observed ratings, and then
use them to predict the ratings for new user-item pairs.
BigQuery ML provides a built-in function called ML.RECOMMEND that can generate
recommendations for a given user based on a trained matrix factorization model. To use
BigQuery ML, you need to load the data in BigQuery, which is a serverless, scalable, and cost-
effective data warehouse. You can use the bq command-line tool, the BigQuery API, or the Cloud
Console to load data from Cloud Storage to BigQuery. Alternatively, you can use federated
queries to query data directly from Cloud Storage without loading it to BigQuery, but this may
incur additional costs and performance overhead. Option A is incorrect because BigQuery ML
does not support Autoencoder models, which are a type of neural network that can learn
compressed representations of the input data. Autoencoder models are not suitable for
recommendation systems, as they do not capture the interactions between users and items.
Option C is incorrect because using TensorFlow to train a two-tower model requires more coding
than using BigQuery ML. A two-tower model is a type of neural network that learns embeddings
for users and items separately, and then combines them with a dot product or a cosine similarity
to compute the rating. TensorFlow is a low-level framework that requires you to define the model
architecture, the loss function, the optimizer, the training loop, and the evaluation metrics.
Moreover, you need to read the data from Cloud Storage to a Vertex AI Workbench notebook,
which is an instance of JupyterLab that runs on a Google Cloud virtual machine. This may involve
additional steps such as authentication, authorization, and data preprocessing. Option D is
incorrect because using TensorFlow to train a matrix factorization model also requires more
coding than using BigQuery ML. Although TensorFlow provides some high-level APIs such as
Keras and TensorFlow Recommenders that can simplify the model development, you still need to
handle the data loading and the model training and evaluation yourself. Furthermore, you need to
read the data from Cloud Storage to a Vertex AI Workbench notebook, which may incur additional
complexity and costs. References:
* BigQuery ML documentation
* Using matrix factorization with BigQuery ML
* Recommendations AI documentation
* Loading data into BigQuery
* Querying data in Cloud Storage from BigQuery
* Vertex AI Workbench documentation
* TensorFlow documentation
* TensorFlow Recommenders documentation

NEW QUESTION: 52
You are developing an ML model in a Vertex Al Workbench notebook. You want to track artifacts
and compare models during experimentation using different approaches. You need to rapidly and
easily transition successful experiments to production as you iterate on your model
implementation. What should you do?
A. 1 Initialize the Vertex SDK with the name of your experiment Log parameters and metrics for
each experiment, and attach dataset and model artifacts as inputs and outputs to each execution.
2 After a successful experiment create a Vertex Al pipeline.
B. 1. Initialize the Vertex SDK with the name of your experiment Log parameters and metrics for
each experiment, save your dataset to a Cloud Storage bucket and upload the models to Vertex
Al Model Registry.
2 After a successful experiment create a Vertex Al pipeline.
C. 1 Create a Vertex Al pipeline with parameters you want to track as arguments to your Pipeline
Job Use the Metrics. Model, and Dataset artifact types from the Kubeflow Pipelines DSL as the
inputs and outputs of the components in your pipeline.
2. Associate the pipeline with your experiment when you submit the job.
D. 1 Create a Vertex Al pipeline Use the Dataset and Model artifact types from the Kubeflow
Pipelines.
DSL as the inputs and outputs of the components in your pipeline.
2. In your training component use the Vertex Al SDK to create an experiment run Configure the
log_params and log_metrics functions to track parameters and metrics of your experiment.
Answer: (SHOW ANSWER)
Vertex AI is a unified platform for building and managing machine learning solutions on Google
Cloud. It provides various services and tools for different stages of the machine learning lifecycle,
such as data preparation, model training, deployment, monitoring, and experimentation. Vertex AI
Workbench is an integrated development environment (IDE) that allows you to create and run
Jupyter notebooks on Google Cloud. You can use Vertex AI Workbench to develop your ML
model in Python, using libraries such as TensorFlow, PyTorch, scikit-learn, etc. You can also use
the Vertex SDK, which is a Python client library for Vertex AI, to track artifacts and compare
models during experimentation. You can use the aiplatform.init function to initialize the Vertex
SDK with the name of your experiment. You can use the aiplatform.start_run and
aiplatform.end_run functions to create and close an experiment run. You can use the
aiplatform.log_params and aiplatform.log_metrics functions to log the parameters and metrics for
each experiment run. You can also use the aiplatform.log_datasets and aiplatform.log_model
functions to attach the dataset and model artifacts as inputs and outputs to each experiment run.
These functions allow you to record and store the metadata and artifacts of your experiments,
and compare them using the Vertex AI Experiments UI. After a successful experiment, you can
create a Vertex AI pipeline, which is a way to automate and orchestrate your ML workflows. You
can use the aiplatform.PipelineJob class to create a pipeline job, and specify the components and
dependencies of your pipeline. You can also use the aiplatform.CustomContainerTrainingJob
class to create a custom container training job, and use the run method to run the job as a
pipeline component. You can use the aiplatform.Model.deploy method to deploy your model as a
pipeline component. You can also use the aiplatform.Model.monitor method to monitor your
model as a pipeline component. By creating a Vertex AI pipeline, you can rapidly and easily
transition successful experiments to production, and reuse and share your ML workflows. This
solution requires minimal changes to your code, and leverages the Vertex AI services and tools to
streamline your ML development process. References: The answer can be verified from official
Google Cloud documentation and resources related to Vertex AI, Vertex AI Workbench, Vertex
SDK, and Vertex AI pipelines.
* Vertex AI | Google Cloud
* Vertex AI Workbench | Google Cloud
* Vertex SDK for Python | Google Cloud
* Vertex AI pipelines | Google Cloud

NEW QUESTION: 53
Your company manages an ecommerce website. You developed an ML model that recommends
additional products to users in near real time based on items currently in the user's cart. The
workflow will include the following processes.
1 The website will send a Pub/Sub message with the relevant data and then receive a message
with the prediction from Pub/Sub.
2 Predictions will be stored in BigQuery
3. The model will be stored in a Cloud Storage bucket and will be updated frequently You want to
minimize prediction latency and the effort required to update the model How should you
reconfigure the architecture?
A. Write a Cloud Function that loads the model into memory for prediction Configure the function
to be triggered when messages are sent to Pub/Sub.
B. Create a pipeline in Vertex Al Pipelines that performs preprocessing, prediction and
postprocessing Configure the pipeline to be triggered by a Cloud Function when messages are
sent to Pub/Sub.
C. Expose the model as a Vertex Al endpoint Write a custom DoFn in a Dataflow job that calls the
endpoint for prediction.
D. Use the Runlnference API with watchFilePatterr. in a Dataflow job that wraps around the
model and serves predictions.
Answer: D (LEAVE A REPLY)
According to the web search results, RunInference API1 is a feature of Apache Beam that
enables you to run models as part of your pipeline in a way that is optimized for machine learning
inference. RunInference API supports features like batching, caching, and model reloading.
RunInference API can be used with various frameworks, such as TensorFlow, PyTorch, Sklearn,
XGBoost, ONNX, and TensorRT1. Dataflow2 is a fully managed service for running Apache
Beam pipelines on Google Cloud. Dataflow handles the provisioning and management of the
compute resources, as well as the optimization and execution of the pipelines. Therefore, option
D is the best way to reconfigure the architecture for the given use case, as it allows you to use the
RunInference API with watchFilePattern in a Dataflow job that wraps around the model and
serves predictions. This way, you can minimize prediction latency and the effort required to
update the model, as the RunInference API will automatically reload the model from the Cloud
Storage bucket whenever there is a change in the model file1. The other options are not relevant
or optimal for this scenario. References:
* RunInference API
* Dataflow
* Google Professional Machine Learning Certification Exam 2023
* Latest Google Professional Machine Learning Engineer Actual Free Exam Questions

NEW QUESTION: 54
You recently trained an XGBoost model on tabular data You plan to expose the model for internal
use as an HTTP microservice After deployment you expect a small number of incoming requests.
You want to productionize the model with the least amount of effort and latency. What should you
do?
A. Deploy the model to BigQuery ML by using CREATE model with the
BOOSTED-THREE-REGRESSOR statement and invoke the BigQuery API from the microservice.
B. Build a Flask-based app Package the app in a custom container on Vertex Al and deploy it to
Vertex Al Endpoints.
C. Build a Flask-based app Package the app in a Docker image and deploy it to Google
Kubernetes Engine in Autopilot mode.
D. Use a prebuilt XGBoost Vertex container to create a model and deploy it to Vertex Al
Endpoints.
Answer: (SHOW ANSWER)
XGBoost is a popular open-source library that provides a scalable and efficient implementation of
gradient boosted trees. You can use XGBoost to train a classification or regression model on
tabular data. You can also use Vertex AI to productionize the model and expose it for internal use
as an HTTP microservice. Vertex AI is a service that allows you to create and train ML models
using Google Cloud technologies. You can use a prebuilt XGBoost Vertex container to create a
model and deploy it to Vertex AI Endpoints. A prebuilt Vertex container is a container image that
contains the dependencies and libraries needed to run a specific ML framework, such as
XGBoost. You can use a prebuilt Vertex container to simplify the model creation and deployment
process, without having to build your own custom container. Vertex AI Endpoints is a service that
allows you to serve your ML models online and scale them automatically. You can use Vertex AI
Endpoints to deploy the model from the prebuilt Vertex container and expose it as an HTTP
microservice. You can also configure the endpoint to handle a small number of incoming
requests, and optimize the latency and cost of serving the model. By using a prebuilt XGBoost
Vertex container and Vertex AI Endpoints, you can productionize the model with the least amount
of effort and latency. References:
* XGBoost documentation
* Vertex AI documentation
* Prebuilt Vertex container documentation
* Vertex AI Endpoints documentation
* Preparing for Google Cloud Certification: Machine Learning Engineer Professional Certificate

NEW QUESTION: 55
You are a lead ML engineer at a retail company. You want to track and manage ML metadata in a
centralized way so that your team can have reproducible experiments by generating artifacts.
Which management solution should you recommend to your team?
A. Store your tf.logging data in BigQuery.
B. Manage all relational entities in the Hive Metastore.
C. Store all ML metadata in Google Cloud's operations suite.
D. Manage your ML workflows with Vertex ML Metadata.
Answer: (SHOW ANSWER)
Vertex ML Metadata is a service that lets you track and manage the metadata produced by your
ML workflows in a centralized way. It helps you have reproducible experiments by generating
artifacts that represent the data, parameters, and metrics used or produced by your ML system.
You can also analyze the lineage and performance of your ML artifacts using Vertex ML
Metadata.
Some of the benefits of using Vertex ML Metadata are:
* It captures your ML system's metadata as a graph, where artifacts and executions are nodes,
and events are edges that link them as inputs or outputs.
* It allows you to create contexts to group sets of artifacts and executions together, such as
experiments, runs, or projects.
* It supports querying and filtering the metadata using the Vertex AI SDK for Python or REST
commands.
* It integrates with other Vertex AI services, such as Vertex AI Pipelines and Vertex AI
Experiments, to automatically log metadata and artifacts.
The other options are not suitable for tracking and managing ML metadata in a centralized way.
* Option A: Storing your tf.logging data in BigQuery is not enough to capture the full metadata of
your ML system, such as the artifacts and their lineage. BigQuery is a data warehouse service
that is mainly used for analytics and reporting, not for metadata management.
* Option B: Managing all relational entities in the Hive Metastore is not a good solution for ML
metadata, as it is designed for storing metadata of Hive tables and partitions, not for ML artifacts
and executions.
Hive Metastore is a component of the Apache Hive project, which is a data warehouse system for
querying and analyzing large datasets stored in Hadoop.
* Option C: Storing all ML metadata in Google Cloud's operations suite is not a feasible option, as
it is a set of tools for monitoring, logging, tracing, and debugging your applications and
infrastructure, not for ML metadata. Google Cloud's operations suite does not provide the
features and integrations that Vertex ML Metadata offers for ML workflows.

NEW QUESTION: 56
Your team is training a large number of ML models that use different algorithms, parameters and
datasets.
Some models are trained in Vertex Ai Pipelines, and some are trained on Vertex Al Workbench
notebook instances. Your team wants to compare the performance of the models across both
services. You want to minimize the effort required to store the parameters and metrics What
should you do?
A. Implement an additional step for all the models running in pipelines and notebooks to export
parameters and metrics to BigQuery.
B. Create a Vertex Al experiment Submit all the pipelines as experiment runs. For models trained
on notebooks log parameters and metrics by using the Vertex Al SDK.
C. Implement all models in Vertex Al Pipelines Create a Vertex Al experiment, and associate all
pipeline runs with that experiment.
D. Store all model parameters and metrics as mode! metadata by using the Vertex Al Metadata
API.
Answer: (SHOW ANSWER)
Vertex AI Experiments is a service that allows you to track, compare, and manage experiments
with Vertex AI. You can use Vertex AI Experiments to record the parameters, metrics, and
artifacts of each model training run, and compare them in a graphical interface. Vertex AI
Experiments supports models trained in Vertex AI Pipelines, Vertex AI Custom Training, and
Vertex AI Workbench notebooks. To use Vertex AI Experiments, you need to create an
experiment and submit your pipeline runs or custom training jobs as experiment runs.
For models trained on notebooks, you need to use the Vertex AI SDK to log the parameters and
metrics to the experiment. This way, you can minimize the effort required to store and compare
the model performance across different services. References: Track, compare, manage
experiments with Vertex AI Experiments, Vertex AI Pipelines: Metrics visualization and run
comparison using the KFP SDK, [Vertex AI SDK for Python]

NEW QUESTION: 57
You are developing an ML model intended to classify whether X-Ray images indicate bone
fracture risk. You have trained on Api Resnet architecture on Vertex AI using a TPU as an
accelerator, however you are unsatisfied with the trainning time and use memory usage. You
want to quickly iterate your training code but make minimal changes to the code. You also want to
minimize impact on the models accuracy. What should you do?
A. Configure your model to use bfloat16 instead float32
B. Reduce the global batch size from 1024 to 256
C. Reduce the number of layers in the model architecture
D. Reduce the dimensions of the images used un the model
Answer: (SHOW ANSWER)
Using bfloat16 instead of float32 can reduce the memory usage and training time of the model,
while having minimal impact on the accuracy. Bfloat16 is a 16-bit floating-point format that
preserves the range of 32-bit floating-point numbers, but reduces the precision from 24 bits to 8
bits. This means that bfloat16 can store the same magnitude of numbers as float32, but with less
detail. Bfloat16 is supported by TPUs and some GPUs, and can be used as a drop-in
replacement for float32 in most cases. Bfloat16 can also improve the numerical stability of the
model, as it reduces the risk of overflow and underflow errors.
Reducing the global batch size, the number of layers, or the dimensions of the images can also
reduce the memory usage and training time of the model, but they can also affect the model's
accuracy and performance.
Reducing the global batch size can make the model less stable and converge slower, as it
reduces the amount of information available for each gradient update. Reducing the number of
layers can make the model less expressive and powerful, as it reduces the depth and complexity
of the network. Reducing the dimensions of the images can make the model less accurate and
robust, as it reduces the resolution and quality of the input data. References:
* Bfloat16: The secret to high performance on Cloud TPUs
* Bfloat16 floating-point format
* How does Batch Size impact your model learning

NEW QUESTION: 58
You are developing a model to identify traffic signs in images extracted from videos taken from
the dashboard of a vehicle. You have a dataset of 100 000 images that were cropped to show
one out of ten different traffic signs. The images have been labeled accordingly for model training
and are stored in a Cloud Storage bucket You need to be able to tune the model during each
training run. How should you train the model?
A. Train a model for object detection by using Vertex Al AutoML.
B. Train a model for image classification by using Vertex Al AutoML.
C. Develop the model training code for object detection and tram a model by using Vertex Al
custom training.
D. Develop the model training code for image classification and train a model by using Vertex Al
custom training.
Answer: (SHOW ANSWER)
Image classification is a task where the model assigns a label to an image based on its content,
such as "stop sign" or "speed limit"1. Object detection is a task where the model locates and
identifies multiple objects in an image, and draws bounding boxes around them2. Since your
dataset consists of images that were cropped to show one out of ten different traffic signs, you
are dealing with an image classification problem, not an object detection problem. Therefore, you
need to train a model for image classification, not object detection.
Vertex AI AutoML is a service that allows you to train and deploy high-quality ML models with
minimal effort and machine learning expertise3. You can use Vertex AI AutoML to train a model
for image classification by uploading your images and labels to a Vertex AI dataset, and then
launching an AutoML training job4. However, Vertex AI AutoML does not allow you to tune the
model during each training run, as it automatically selects the best model architecture and
hyperparameters for your data4.
Vertex AI custom training is a service that allows you to train and deploy your own custom ML
models using your own code and frameworks5. You can use Vertex AI custom training to train a
model for image classification by writing your own model training code, such as using TensorFlow
or PyTorch, and then creating and running a custom training job. Vertex AI custom training allows
you to tune the model during each training run, as you can specify the model architecture and
hyperparameters in your code, and use Vertex AI Hyperparameter Tuning to optimize them .
Therefore, the best option for your scenario is to develop the model training code for image
classification and train a model by using Vertex AI custom training.
References:
* Image classification | TensorFlow Core
* Object detection | TensorFlow Core
* Introduction to Vertex AI AutoML | Google Cloud
* AutoML Vision | Google Cloud
* Introduction to Vertex AI custom training | Google Cloud
* [Custom training with TensorFlow | Vertex AI | Google Cloud]
* [Hyperparameter tuning overview | Vertex AI | Google Cloud]

NEW QUESTION: 59
You trained a text classification model. You have the following SignatureDefs:

What is the correct way to write the predict request?


A. data = json.dumps({"signature_name": "serving_default'\ "instances": [fab', 'be1, 'cd']]})
B. data = json dumps({"signature_name": "serving_default"! "instances": [['a', 'b', "c", 'd', 'e', 'f']]})
C. data = json.dumps({"signature_name": "serving_default, "instances": [['a', 'b\ 'c'1, [d\ 'e\ T]]})
D. data = json dumps({"signature_name": f,serving_default", "instances": [['a', 'b'], [c\ 'd'], ['e\ T]]})
Answer: (SHOW ANSWER)
A predict request is a way to send data to a trained model and get predictions in return. A predict
request can be written in different formats, such as JSON, protobuf, or gRPC, depending on the
service and the platform that are used to host and serve the model. A predict request usually
contains the following information:
* The signature name: This is the name of the signature that defines the inputs and outputs of the
model. A signature is a way to specify the expected format, type, and shape of the data that the
model can accept and produce. A signature can be specified when exporting or saving the model,
or it can be automatically inferred by the service or the platform. A model can have multiple
signatures, but only one can be used for each predict request.
* The instances: This is the data that is sent to the model for prediction. The instances can be a
single instance or a batch of instances, depending on the size and shape of the data. The
instances should match the input specification of the signature, such as the number, name, and
type of the input tensors.
For the use case of training a text classification model, the correct way to write the predict request
is D. data = json.dumps({"signature_name": "serving_default", "instances": [['a', 'b'], ['c', 'd'], ['e',
'f']]}) This option involves writing the predict request in JSON format, which is a common and
convenient format for sending and receiving data over the web. JSON stands for JavaScript
Object Notation, and it is a way to represent data as a collection of name-value pairs or an
ordered list of values. JSON can be easily converted to and from Python objects using the json
module.
This option also involves using the signature name "serving_default", which is the default
signature name that is assigned to the model when it is saved or exported without specifying a
custom signature name. The serving_default signature defines the input and output tensors of the
model based on the SignatureDef that is shown in the image. According to the SignatureDef, the
model expects an input tensor called "text" that has a shape of (-1, 2) and a type of DT_STRING,
and produces an output tensor called "softmax" that has a shape of (-1, 2) and a type of
DT_FLOAT. The -1 in the shape indicates that the dimension can vary depending on the number
of instances, and the 2 indicates that the dimension is fixed at 2. The DT_STRING and
DT_FLOAT indicate that the data type is string and float, respectively.
This option also involves sending a batch of three instances to the model for prediction. Each
instance is a list of two strings, such as ['a', 'b'], ['c', 'd'], or ['e', 'f']. These instances match the
input specification of the signature, as they have a shape of (3, 2) and a type of string. The model
will process these instances and produce a batch of three predictions, each with a softmax output
that has a shape of (1, 2) and a type of float.
The softmax output is a probability distribution over the two possible classes that the model can
predict, such as positive or negative sentiment.
Therefore, writing the predict request as data = json.dumps({"signature_name": "serving_default",
"instances": [['a', 'b'], ['c', 'd'], ['e', 'f']]}) is the correct and valid way to send data to the text
classification model and get predictions in return.
References:
* [json - JSON encoder and decoder]

NEW QUESTION: 60

A. Use the BigQuery client library to load data into a dataframe, and use
tf.data.Dataset.from_tensor_slices() to read it.
B. Export data to CSV files in Cloud Storage, and use tf.data.TextLineDataset() to read them.
C. Convert the data into TFRecords, and use tf.data.TFRecordDataset() to read them.
D. Use TensorFlow I/O's BigQuery Reader to directly read the data.
Answer: (SHOW ANSWER)
The best option for developing and comparing multiple models on a large-scale BigQuery table
using TensorFlow and Vertex AI is to use TensorFlow I/O's BigQuery Reader to directly read the
data. This option has the following advantages:
* It minimizes any bottlenecks during the data ingestion stage, as the BigQuery Reader can
stream data from BigQuery to TensorFlow in parallel and in batches, without loading the entire
table into memory or disk. The BigQuery Reader can also perform data transformations and
filtering using SQL queries, reducing the need for additional preprocessing steps in TensorFlow.
* It leverages the scalability and performance of BigQuery, as the BigQuery Reader can handle
hundreds of millions of records worth of training data efficiently and reliably. BigQuery is a
serverless, fully managed, and highly scalable data warehouse that can run complex queries over
petabytes of data in seconds.
* It simplifies the integration with Vertex AI, as the BigQuery Reader can be used with both
custom and pre-built TensorFlow models on Vertex AI. Vertex AI is a unified platform for machine
learning that provides various tools and features for data ingestion, data labeling, data
preprocessing, model training, model tuning, model deployment, model monitoring, and model
explainability.
The other options are less optimal for the following reasons:
* Option A: Using the BigQuery client library to load data into a dataframe, and using
tf.data.Dataset.from_tensor_slices() to read it, introduces memory and performance issues. This
option requires loading the entire BigQuery table into a Pandas dataframe, which can consume a
lot of memory and cause out-of-memory errors. Moreover, using
tf.data.Dataset.from_tensor_slices() to read the dataframe can be slow and inefficient, as it
creates one slice per row of the dataframe, resulting in a large number of small tensors.
* Option B: Exporting data to CSV files in Cloud Storage, and using tf.data.TextLineDataset() to
read them, introduces additional steps and complexity. This option requires exporting the
BigQuery table to one or more CSV files in Cloud Storage, which can take a long time and
consume a lot of storage space.
Moreover, using tf.data.TextLineDataset() to read the CSV files can be slow and error-prone, as it
requires parsing and decoding each line of text, handling missing values and invalid data, and
applying data transformations and validations.
* Option C: Converting the data into TFRecords, and using tf.data.TFRecordDataset() to read
them, introduces additional steps and complexity. This option requires converting the BigQuery
table into one or more TFRecord files, which are binary files that store serialized TensorFlow
examples. This can take a long time and consume a lot of storage space. Moreover, using
tf.data.TFRecordDataset() to read the TFRecord files requires defining and parsing the schema of
the TensorFlow examples, which can be tedious and error-prone.
References:
* [TensorFlow I/O documentation]
* [BigQuery documentation]
* [Vertex AI documentation]
NEW QUESTION: 61
You are designing an architecture with a serverless ML system to enrich customer support tickets
with informative metadata before they are routed to a support agent. You need a set of models to
predict ticket priority, predict ticket resolution time, and perform sentiment analysis to help agents
make strategic decisions when they process support requests. Tickets are not expected to have
any domain-specific terms or jargon.
The proposed architecture has the following flow:

Which endpoints should the Enrichment Cloud Functions call?


A. 1 = Vertex Al. 2 = Vertex Al. 3 = AutoML Natural Language
B. 1 = Vertex Al. 2 = Vertex Al. 3 = Cloud Natural Language API
C. 1 = Vertex Al. 2 = Vertex Al. 3 = AutoML Vision
D. 1 = Cloud Natural Language API. 2 = Vertex Al, 3 = Cloud Vision API
Answer: (SHOW ANSWER)
Vertex AI is a unified platform for building and deploying ML models on Google Cloud. It supports
both custom and AutoML models, and provides various tools and services for ML development,
such as Vertex Pipelines, Vertex Vizier, Vertex Explainable AI, and Vertex Feature Store. Vertex
AI can be used to create models for predicting ticket priority and resolution time, as these are
domain-specific tasks that require custom training data and evaluation metrics. Cloud Natural
Language API is a pre-trained service that provides natural language understanding capabilities,
such as sentiment analysis, entity analysis, syntax analysis, and content classification. Cloud
Natural Language API can be used to perform sentiment analysis on the support tickets, as this is
a general task that does not require domain-specific knowledge or jargon. The other options are
not suitable for the given architecture. AutoML Natural Language and AutoML Vision are services
that allow users to create custom natural language and vision models using their own data and
labels. They are not needed for sentiment analysis, as Cloud Natural Language API already
provides this functionality. Cloud Vision API is a pre-trained service that provides image analysis
capabilities, such as object detection, face detection, text detection, and image labeling. It is not
relevant for the support tickets, as they are not expected to have any images. References:
* Vertex AI documentation
* Cloud Natural Language API documentation

Valid Professional-Machine-Learning-Engineer Dumps shared by ExamDiscuss.com for


Helping Passing Professional-Machine-Learning-Engineer Exam! ExamDiscuss.com now offer
the newest Professional-Machine-Learning-Engineer exam dumps, the ExamDiscuss.com
Professional-Machine-Learning-Engineer exam questions have been updated and answers
have been corrected get the newest ExamDiscuss.com Professional-Machine-Learning-
Engineer dumps with Test Engine here:
https://www.examdiscuss.com/Google/exam/Professional-Machine-Learning-
Engineer/premium/ (273 Q&As Dumps, 35%OFF Special Discount Code: freecram)

NEW QUESTION: 62
You have built a custom model that performs several memory-intensive preprocessing tasks
before it makes a prediction. You deployed the model to a Vertex Al endpoint. and validated that
results were received in a reasonable amount of time After routing user traffic to the endpoint, you
discover that the endpoint does not autoscale as expected when receiving multiple requests What
should you do?
A. Use a machine type with more memory
B. Decrease the number of workers per machine
C. Increase the CPU utilization target in the autoscaling configurations
D. Decrease the CPU utilization target in the autoscaling configurations
Answer: (SHOW ANSWER)
According to the web search results, Vertex AI is a unified platform for machine learning
development and deployment. Vertex AI offers various services and tools for building, managing,
and serving machine learning models1. Vertex AI allows you to deploy your models to endpoints
for online prediction, and configure the compute resources and autoscaling options for your
deployed models2. Autoscaling with Vertex AI endpoints is (by default) based on the CPU
utilization across all cores of the machine type you have specified. The default threshold of 60%
represents 60% on all cores. For example, for a 4 core machine, that means you need
240% utilization to trigger autoscaling3. Therefore, if you discover that the endpoint does not
autoscale as expected when receiving multiple requests, you might need to decrease the CPU
utilization target in the autoscaling configurations. This way, you can lower the threshold for
triggering autoscaling and allocate more resources to handle the prediction requests. Therefore,
option D is the best way to solve the problem for the given use case. The other options are not
relevant or optimal for this scenario. References:
* Vertex AI
* Deploy a model to an endpoint
* Vertex AI endpoint doesn't scale up / down
* Google Professional Machine Learning Certification Exam 2023
* Latest Google Professional Machine Learning Engineer Actual Free Exam Questions

NEW QUESTION: 63
You are going to train a DNN regression model with Keras APIs using this code:
How many trainable weights does your model have? (The arithmetic below is correct.)
A. 501*256+257*128+2 = 161154
B. 500*256+256*128+128*2 = 161024
C. 501*256+257*128+128*2=161408
D. 500*256*0 25+256*128*0 25+128*2 = 40448
Answer: (SHOW ANSWER)
The number of trainable weights in a DNN regression model with Keras APIs can be calculated
by multiplying the number of input units by the number of output units for each layer, and adding
the number of bias units for each layer. The bias units are usually equal to the number of output
units, except for the last layer, which does not have bias units if the activation function is
softmax1. In this code, the model has three layers: a dense layer with 256 units and relu
activation, a dropout layer with 0.25 rate, and a dense layer with 2 units and softmax activation.
The input shape is 500. Therefore, the number of trainable weights is:
* For the first layer: 500 input units * 256 output units + 256 bias units = 128256
* For the second layer: The dropout layer does not have any trainable weights, as it only
randomly sets some of the input units to zero to prevent overfitting2.
* For the third layer: 256 input units * 2 output units + 0 bias units = 512 The total number of
trainable weights is 128256 + 512 = 161024. Therefore, the correct answer is B.
References:
* How to calculate the number of parameters for a Convolutional Neural Network?
* Dropout (keras.io)

NEW QUESTION: 64
You are training a Resnet model on Al Platform using TPUs to visually categorize types of defects
in automobile engines. You capture the training profile using the Cloud TPU profiler plugin and
observe that it is highly input-bound. You want to reduce the bottleneck and speed up your model
training process. Which modifications should you make to the tf .data dataset?
Choose 2 answers
A. Use the interleave option for reading data
B. Reduce the value of the repeat parameter
C. Increase the buffer size for the shuffle option.
D. Set the prefetch option equal to the training batch size
E. Decrease the batch size argument in your transformation
Answer: (SHOW ANSWER)
The tf.data dataset is a TensorFlow API that provides a way to create and manipulate data
pipelines for machine learning. The tf.data dataset allows you to apply various transformations to
the data, such as reading, shuffling, batching, prefetching, and interleaving. These
transformations can affect the performance and efficiency of the model training process1 One of
the common performance issues in model training is input-bound, which means that the model is
waiting for the input data to be ready and is not fully utilizing the computational resources. Input-
bound can be caused by slow data loading, insufficient parallelism, or large data size. Input-
bound can be detected by using the Cloud TPU profiler plugin, which is a tool that helps you
analyze the performance of your model on Cloud TPUs. The Cloud TPU profiler plugin can show
you the percentage of time that the TPU cores are idle, which indicates input-bound2 To reduce
the input-bound bottleneck and speed up the model training process, you can make some
modifications to the tf.data dataset. Two of the modifications that can help are:
* Use the interleave option for reading data. The interleave option allows you to read data from
multiple files in parallel and interleave their records. This can improve the data loading speed and
reduce the idle time of the TPU cores. The interleave option can be applied by using the
tf.data.Dataset.interleave method, which takes a function that returns a dataset for each input
element, and a number of parallel calls3
* Set the prefetch option equal to the training batch size. The prefetch option allows you to
prefetch the next batch of data while the current batch is being processed by the model. This can
reduce the latency
* between batches and improve the throughput of the model training. The prefetch option can be
applied by using the tf.data.Dataset.prefetch method, which takes a buffer size argument. The
buffer size should be equal to the training batch size, which is the number of examples per batch4
The other options are not effective or counterproductive. Reducing the value of the repeat
parameter will reduce the number of epochs, which is the number of times the model sees the
entire dataset. This can affect the model's accuracy and convergence. Increasing the buffer size
for the shuffle option will increase the randomness of the data, but also increase the memory
usage and the data loading time. Decreasing the batch size argument in your transformation will
reduce the number of examples per batch, which can affect the model's stability and
performance.
References: 1: tf.data: Build TensorFlow input pipelines 2: Cloud TPU Tools in TensorBoard 3:
tf.data.Dataset.interleave 4: tf.data.Dataset.prefetch : [Better performance with the tf.data API]

NEW QUESTION: 65
You work for a company that is developing a new video streaming platform. You have been asked
to create a recommendation system that will suggest the next video for a user to watch. After a
review by an AI Ethics team, you are approved to start development. Each video asset in your
company's catalog has useful metadata (e.g., content type, release date, country), but you do not
have any historical user event data. How should you build the recommendation system for the
first version of the product?
A. Launch the product without machine learning. Present videos to users alphabetically, and start
collecting user event data so you can develop a recommender model in the future.
B. Launch the product without machine learning. Use simple heuristics based on content
metadata to recommend similar videos to users, and start collecting user event data so you can
develop a recommender model in the future.
C. Launch the product with machine learning. Use a publicly available dataset such as MovieLens
to train a model using the Recommendations AI, and then apply this trained model to your data.
D. Launch the product with machine learning. Generate embeddings for each video by training an
autoencoder on the content metadata using TensorFlow. Cluster content based on the similarity
of these embeddings, and then recommend videos from the same cluster.
Answer: (SHOW ANSWER)
The best option for building a recommendation system without any user event data is to use
simple heuristics based on content metadata. This is a type of content-based filtering, which
recommends items that are similar to the ones that the user has interacted with or selected,
based on their attributes. For example, if a user selects a comedy movie from the US released in
2020, the system can recommend other comedy movies from the US released in 2020 or nearby
years. This approach does not require any machine learning, but it can leverage the existing
metadata of the videos to provide relevant recommendations. It also allows the system to start
collecting user event data, such as views, likes, ratings, etc., which can be used to train a more
sophisticated machine learning model in the future, such as a collaborative filtering model or a
hybrid model that combines content and collaborative information. References:
* Recommendation Systems
* Content-Based Filtering
* Collaborative Filtering
* Hybrid Recommender Systems: A Systematic Literature Review

NEW QUESTION: 66
You lead a data science team at a large international corporation. Most of the models your team
trains are large-scale models using high-level TensorFlow APIs on AI Platform with GPUs. Your
team usually takes a few weeks or months to iterate on a new version of a model. You were
recently asked to review your team's spending. How should you reduce your Google Cloud
compute costs without impacting the model's performance?
A. Use AI Platform to run distributed training jobs with checkpoints.
B. Use AI Platform to run distributed training jobs without checkpoints.
C. Migrate to training with Kuberflow on Google Kubernetes Engine, and use preemptible VMs
with checkpoints.
D. Migrate to training with Kuberflow on Google Kubernetes Engine, and use preemptible VMs
without checkpoints.
Answer: (SHOW ANSWER)
* Option A is incorrect because using AI Platform to run distributed training jobs with checkpoints
does not reduce the compute costs, but rather increases them by using more resources and
storing the checkpoints.
* Option B is incorrect because using AI Platform to run distributed training jobs without
checkpoints may reduce the compute costs, but it also risks losing the progress of the training if
the job fails or is interrupted.
* Option C is correct because migrating to training with Kubeflow on Google Kubernetes Engine,
and using preemptible VMs with checkpoints can reduce the compute costs significantly by using
cheaper and more scalable resources, while also preserving the state of the training with
checkpoints.
* Option D is incorrect because using preemptible VMs without checkpoints may reduce the
compute costs, but it also risks losing the training progress if the VMs are preempted.
References:
* Kubeflow on Google Cloud
* Using preemptible VMs and GPUs
* Saving and loading models

NEW QUESTION: 67
You work for an auto insurance company. You are preparing a proof-of-concept ML application
that uses images of damaged vehicles to infer damaged parts Your team has assembled a set of
annotated images from damage claim documents in the company's database The annotations
associated with each image consist of a bounding box for each identified damaged part and the
part name. You have been given a sufficient budget to tram models on Google Cloud You need to
quickly create an initial model What should you do?
A. Download a pre-trained object detection mode! from TensorFlow Hub Fine-tune the model in
Vertex Al Workbench by using the annotated image data.
B. Train an object detection model in AutoML by using the annotated image data.
C. Create a pipeline in Vertex Al Pipelines and configure the AutoMLTrainingJobRunOp compon
it to train a custom object detection model by using the annotated image data.
D. Train an object detection model in Vertex Al custom training by using the annotated image
data.
Answer: (SHOW ANSWER)
According to the official exam guide1, one of the skills assessed in the exam is to "design, build,
and productionalize ML models to solve business challenges using Google Cloud technologies".
AutoML Vision2 is a service that allows you to train and deploy custom vision models for image
classification and object detection. AutoML Vision simplifies the model development process by
providing a graphical user interface and a no-code approach. You can use AutoML Vision to train
an object detection model by using the annotated image data, and evaluate the model
performance using metrics such as mean average precision (mAP) and intersection over union
(IoU)3. Therefore, option B is the best way to quickly create an initial model for the given use
case. The other options are not relevant or optimal for this scenario. References:
* Professional ML Engineer Exam Guide
* AutoML Vision
* Object detection evaluation
* Google Professional Machine Learning Certification Exam 2023
* Latest Google Professional Machine Learning Engineer Actual Free Exam Questions

NEW QUESTION: 68
You developed an ML model with Al Platform, and you want to move it to production. You serve a
few thousand queries per second and are experiencing latency issues. Incoming requests are
served by a load balancer that distributes them across multiple Kubeflow CPU-only pods running
on Google Kubernetes Engine (GKE). Your goal is to improve the serving latency without
changing the underlying infrastructure.
What should you do?
A. Significantly increase the max_batch_size TensorFlow Serving parameter
B. Switch to the tensorflow-model-server-universal version of TensorFlow Serving
C. Significantly increase the max_enqueued_batches TensorFlow Serving parameter
D. Recompile TensorFlow Serving using the source to support CPU-specific optimizations Instruct
GKE to choose an appropriate baseline minimum CPU platform for serving nodes
Answer: (SHOW ANSWER)
TensorFlow Serving is a service that allows you to deploy and serve TensorFlow models in a
scalable and efficient way. TensorFlow Serving supports various platforms and hardware, such as
CPU, GPU, and TPU.
However, the default TensorFlow Serving binaries are built with generic CPU instructions, which
may not leverage the full potential of the CPU architecture. To improve the serving latency and
performance, you can recompile TensorFlow Serving using the source code and enable CPU-
specific optimizations, such as AVX, AVX2, and FMA1. These optimizations can speed up the
computation and inference of the TensorFlow models, especially for deep neural networks.
Google Kubernetes Engine (GKE) is a service that allows you to run and manage containerized
applications on Google Cloud using Kubernetes. GKE supports various types and sizes of nodes,
which are the virtual machines that run the containers. GKE also supports different CPU
platforms, which are the generations and models of the CPUs that power the nodes. GKE allows
you to choose a baseline minimum CPU platform for your node pool, which is a group of nodes
with the same configuration. By choosing a baseline minimum CPU platform, you can ensure that
your nodes have the CPU features and capabilities that match your workload requirements2.
For the use case of serving a few thousand queries per second and experiencing latency issues,
the best option is to recompile TensorFlow Serving using the source to support CPU-specific
optimizations, and instruct GKE to choose an appropriate baseline minimum CPU platform for
serving nodes. This option can improve the serving latency and performance without changing
the underlying infrastructure, as it only involves rebuilding the TensorFlow Serving binary and
selecting the CPU platform for the GKE nodes. This option can also take advantage of the CPU-
only pods that are running on GKE, as it can optimize the CPU utilization and efficiency.
Therefore, recompiling TensorFlow Serving using the source to support CPU-specific
optimizations and instructing GKE to choose an appropriate baseline minimum CPU platform for
serving nodes is the best option for this use case.
References:
* Building TensorFlow Serving from source
* Specifying a minimum CPU platform for a node pool

NEW QUESTION: 69
You work for a pharmaceutical company based in Canada. Your team developed a BigQuery ML
model to predict the number of flu infections for the next month in Canada Weather data is
published weekly and flu infection statistics are published monthly. You need to configure a model
retraining policy that minimizes cost What should you do?
A. Download the weather and flu data each week Configure Cloud Scheduler to execute a Vertex
Al pipeline to retrain the model weekly.
B. Download the weather and flu data each month Configure Cloud Scheduler to execute a
Vertex Al pipeline to retrain the model monthly.
C. Download the weather and flu data each week Configure Cloud Scheduler to execute a Vertex
Al pipeline to retrain the model every month.
D. Download the weather data each week, and download the flu data each month Deploy the
model to a Vertex Al endpoint with feature drift monitoring. and retrain the model if a monitoring
alert is detected.
Answer: (SHOW ANSWER)
To configure a model retraining policy that minimizes cost, you should follow these steps:
* Download the weather data each week, and download the flu data each month. This way, you
can keep your data up to date with the latest information available, without downloading
unnecessary or redundant data.
* Deploy the model to a Vertex AI endpoint with feature drift monitoring. This feature allows you to
detect when the distribution of the input data changes significantly from the training data, which
could affect the model performance1.
* Retrain the model if a monitoring alert is detected. This way, you can update your model only
when needed, instead of retraining it on a fixed schedule, which could incur more cost and time.
References:
* 1: Monitor models for feature drift | Vertex AI | Google Cloud

NEW QUESTION: 70

A. Increase the dropout rate to 0.8 in_PREDICT mode by adjusting the TensorFlow Serving
parameters
B. Increase the dropout rate to 0.8 and retrain your model.
C. Switch from CPU to GPU serving
D. Apply quantization to your SavedModel by reducing the floating point precision to tf.float16.
Answer: (SHOW ANSWER)
* Quantization is a technique that reduces the numerical precision of the weights and activations
of a neural network, which can improve the inference speed and reduce the memory footprint of
the model1.
* Reducing the floating point precision from tf.float64 to tf.float16 can potentially halve the latency
and
* memory usage of the model, while having minimal impact on the accuracy2.
* Increasing the dropout rate to 0.8 in either mode would not affect the latency, but would likely
degrade the performance of the model significantly, as dropout is a regularization technique that
randomly drops out units during training to prevent overfitting3.
* Switching from CPU to GPU serving may or may not improve the latency, depending on the
hardware specifications and the model complexity, but it would also incur additional costs and
complexity for deployment4

NEW QUESTION: 71
You work for a magazine publisher and have been tasked with predicting whether customers will
cancel their annual subscription. In your exploratory data analysis, you find that 90% of
individuals renew their subscription every year, and only 10% of individuals cancel their
subscription. After training a NN Classifier, your model predicts those who cancel their
subscription with 99% accuracy and predicts those who renew their subscription with 82%
accuracy. How should you interpret these results?
A. This is not a good result because the model should have a higher accuracy for those who
renew their subscription than for those who cancel their subscription.
B. This is not a good result because the model is performing worse than predicting that people
will always renew their subscription.
C. This is a good result because predicting those who cancel their subscription is more difficult,
since there is less data for this group.
D. This is a good result because the accuracy across both groups is greater than 80%.
Answer: (SHOW ANSWER)
This is not a good result because the model is performing worse than predicting that people will
always renew their subscription. This option has the following reasons:
* It indicates that the model is not learning from the data, but rather memorizing the majority
class. Since
90% of the individuals renew their subscription every year, the model can achieve a 90%
accuracy by simply predicting that everyone will renew their subscription, without considering the
features or the patterns in the data. However, the model's accuracy for predicting those who
renew their subscription is
* only 82%, which is lower than the baseline accuracy of 90%. This suggests that the model is
overfitting to the minority class (those who cancel their subscription), and underfitting to the
majority class (those who renew their subscription).
* It implies that the model is not useful for the business problem, as it cannot identify the
customers who are at risk of churning. The goal of predicting whether customers will cancel their
annual subscription is to prevent customer churn and increase customer retention. However, the
model's accuracy for predicting those who cancel their subscription is 99%, which is too high and
unrealistic, as it means that the model can almost perfectly identify the customers who will churn,
without any false positives or false negatives. This may indicate that the model is cheating or
exploiting some leakage in the data, such as a feature that reveals the outcome of the prediction.
Moreover, the model's accuracy for predicting those who renew their subscription is 82%, which
is too low and unreliable, as it means that the model can miss many customers who will churn,
and falsely label them as renewing customers. This can lead to losing customers and revenue,
and failing to take proactive actions to retain them.
References:
* How to Evaluate Machine Learning Models: Classification Metrics | Machine Learning Mastery
* Imbalanced Classification: Predicting Subscription Churn | Machine Learning Mastery

NEW QUESTION: 72
You work for a company that captures live video footage of checkout areas in their retail stores
You need to use the live video footage to build a mode! to detect the number of customers waiting
for service in near real time You want to implement a solution quickly and with minimal effort How
should you build the model?
A. Use the Vertex Al Vision Occupancy Analytics model.
B. Use the Vertex Al Vision Person/vehicle detector model
C. Train an AutoML object detection model on an annotated dataset by using Vertex AutoML
D. Train a Seq2Seq+ object detection model on an annotated dataset by using Vertex AutoML
Answer: (SHOW ANSWER)
According to the official exam guide1, one of the skills assessed in the exam is to "design, build,
and productionalize ML models to solve business challenges using Google Cloud technologies".
The Vertex AI Vision Occupancy Analytics model2 is a specialized pre-built vision model that lets
you count people or vehicles given specific inputs you add in video frames. It provides advanced
features such as active zones counting, line crossing counting, and dwelling detection. This
model is suitable for the use case of detecting the number of customers waiting for service in near
real time. You can easily create and deploy an occupancy analytics application using Vertex AI
Vision3. The other options are not relevant or optimal for this scenario.
References:
* Professional ML Engineer Exam Guide
* Occupancy analytics guide
* Create an occupancy analytics app with BigQuery forecasting
* Google Professional Machine Learning Certification Exam 2023
* Latest Google Professional Machine Learning Engineer Actual Free Exam Questions

NEW QUESTION: 73
You work for the AI team of an automobile company, and you are developing a visual defect
detection model using TensorFlow and Keras. To improve your model performance, you want to
incorporate some image augmentation functions such as translation, cropping, and contrast
tweaking. You randomly apply these functions to each training batch. You want to optimize your
data processing pipeline for run time and compute resources utilization. What should you do?
A. Embed the augmentation functions dynamically in the tf.Data pipeline.
B. Embed the augmentation functions dynamically as part of Keras generators.
C. Use Dataflow to create all possible augmentations, and store them as TFRecords.
D. Use Dataflow to create the augmentations dynamically per training run, and stage them as
TFRecords.
Answer: (SHOW ANSWER)
The best option for optimizing the data processing pipeline for run time and compute resources
utilization is to embed the augmentation functions dynamically in the tf.Data pipeline. This option
has the following advantages:
* It allows the data augmentation to be performed on the fly, without creating or storing additional
copies of the data. This saves storage space and reduces the data transfer time.
* It leverages the parallelism and performance of the tf.Data API, which can efficiently apply the
augmentation functions to multiple batches of data in parallel, using multiple CPU cores or GPU
* devices. The tf.Data API also supports various optimization techniques, such as caching,
prefetching, and autotuning, to improve the data processing speed and reduce the latency.
* It integrates seamlessly with the TensorFlow and Keras models, which can consume the tf.Data
datasets as inputs for training and evaluation. The tf.Data API also supports various data formats,
such as images, text, audio, and video, and various data sources, such as files, databases, and
web services.
The other options are less optimal for the following reasons:
* Option B: Embedding the augmentation functions dynamically as part of Keras generators
introduces some limitations and overhead. Keras generators are Python generators that yield
batches of data for training or evaluation. However, Keras generators are not compatible with the
tf.distribute API, which is used to distribute the training across multiple devices or machines.
Moreover, Keras generators are not as efficient or scalable as the tf.Data API, as they run on a
single Python thread and do not support parallelism or optimization techniques.
* Option C: Using Dataflow to create all possible augmentations, and store them as TFRecords
introduces additional complexity and cost. Dataflow is a fully managed service that runs Apache
Beam pipelines for data processing and transformation. However, using Dataflow to create all
possible augmentations requires generating and storing a large number of augmented images,
which can consume a lot of storage space and incur storage and network costs. Moreover, using
Dataflow to create the augmentations requires writing and deploying a separate Dataflow
pipeline, which can be tedious and time-consuming.
* Option D: Using Dataflow to create the augmentations dynamically per training run, and stage
them as TFRecords introduces additional complexity and latency. Dataflow is a fully managed
service that runs Apache Beam pipelines for data processing and transformation. However, using
Dataflow to create the augmentations dynamically per training run requires running a Dataflow
pipeline every time the model is trained, which can introduce latency and delay the training
process. Moreover, using Dataflow to create the augmentations requires writing and deploying a
separate Dataflow pipeline, which can be tedious and time-consuming.
References:
* [tf.data: Build TensorFlow input pipelines]
* [Image augmentation | TensorFlow Core]
* [Dataflow documentation]

NEW QUESTION: 74
Your team trained and tested a DNN regression model with good results. Six months after
deployment, the model is performing poorly due to a change in the distribution of the input data.
How should you address the input differences in production?
A. Create alerts to monitor for skew, and retrain the model.
B. Perform feature selection on the model, and retrain the model with fewer features
C. Retrain the model, and select an L2 regularization parameter with a hyperparameter tuning
service
D. Perform feature selection on the model, and retrain the model on a monthly basis with fewer
features
Answer: (SHOW ANSWER)
The performance of a DNN regression model can degrade over time due to a change in the
distribution of the input data. This phenomenon is known as data drift or concept drift, and it can
affect the accuracy and reliability of the model predictions. Data drift can be caused by various
factors, such as seasonal changes, population shifts, market trends, or external events1 To
address the input differences in production, one should create alerts to monitor for skew, and
retrain the model. Skew is a measure of how much the input data in production differs from the
input data used for training the model. Skew can be detected by comparing the statistics and
distributions of the input features in the training and production data, such as mean, standard
deviation, histogram, or quantiles. Alerts can be set up to notify the model developers or
operators when the skew exceeds a certain threshold, indicating a significant change in the input
data2 When an alert is triggered, the model should be retrained with the latest data that reflects
the current distribution of the input features. Retraining the model can help the model adapt to the
new data and improve its performance. Retraining the model can be done manually or
automatically, depending on the frequency and severity of the data drift. Retraining the model can
also involve updating the model architecture, hyperparameters, or optimization algorithm, if
necessary3 The other options are not as effective or feasible. Performing feature selection on the
model and retraining the model with fewer features is not a good idea, as it may reduce the
expressiveness and complexity of the model, and ignore some important features that may affect
the output. Retraining the model and selecting an L2 regularization parameter with a
hyperparameter tuning service is not relevant, as L2 regularization is a technique to prevent
overfitting, not data drift. Retraining the model on a monthly basis with fewer features is not
optimal, as it may not capture the timely changes in the input data, and may compromise the
model performance.
References: 1: Data drift detection for machine learning models 2: Skew and drift detection 3:
Retraining machine learning models

NEW QUESTION: 75
You work for a retail company. You have created a Vertex Al forecast model that produces
monthly item sales predictions. You want to quickly create a report that will help to explain how
the model calculates the predictions. You have one month of recent actual sales data that was
not included in the training dataset. How should you generate data for your report?
A. Create a batch prediction job by using the actual sales data Compare the predictions to the
actuals in the report.
B. Create a batch prediction job by using the actual sates data and configure the job settings to
generate feature attributions. Compare the results in the report.
C. Generate counterfactual examples by using the actual sales data Create a batch prediction job
using the actual sales data and the counterfactual examples Compare the results in the report.
D. Train another model by using the same training dataset as the original and exclude some
columns. Using the actual sales data create one batch prediction job by using the new model and
another one with the original model Compare the two sets of predictions in the report.
Answer: (SHOW ANSWER)
According to the official exam guide1, one of the skills assessed in the exam is to "explain the
predictions of a trained model". Vertex AI provides feature attributions using Shapley Values, a
cooperative game theory algorithm that assigns credit to each feature in a model for a particular
outcome2. Feature attributions can help you understand how the model calculates the predictions
and debug or optimize the model accordingly. You can use Forecasting with AutoML or Tabular
Workflow for Forecasting to generate and query local feature attributions2. The other options are
not relevant or optimal for this scenario. References:
* Professional ML Engineer Exam Guide
* Feature attributions for forecasting
* Google Professional Machine Learning Certification Exam 2023
* Latest Google Professional Machine Learning Engineer Actual Free Exam Questions

NEW QUESTION: 76
You work for a large technology company that wants to modernize their contact center. You have
been asked to develop a solution to classify incoming calls by product so that requests can be
more quickly routed to the correct support team. You have already transcribed the calls using the
Speech-to-Text API. You want to minimize data preprocessing and development time. How
should you build the model?
A. Use the Al Platform Training built-in algorithms to create a custom model
B. Use AutoML Natural Language to extract custom entities for classification
C. Use the Cloud Natural Language API to extract custom entities for classification
D. Build a custom model to identify the product keywords from the transcribed calls, and then run
the keywords through a classification algorithm
Answer: (SHOW ANSWER)
AutoML Natural Language is a service that allows users to create custom natural language
models using their own data and labels. It supports various natural language tasks, such as text
classification, entity extraction, and sentiment analysis. AutoML Natural Language can be used to
build a model to classify incoming calls by product, as it can extract custom entities from the
transcribed calls and assign them to predefined categories.
AutoML Natural Language also minimizes data preprocessing and development time, as it
handles the data preparation, model training, and evaluation automatically. The other options are
not as suitable for this scenario. AI Platform Training built-in algorithms are a set of pre-defined
algorithms that can be used to train ML models on AI Platform, but they do not support natural
language processing tasks. Cloud Natural Language API is a pre-trained service that provides
natural language understanding capabilities, such as sentiment analysis, entity analysis, syntax
analysis, and content classification. However, it does not support custom entities or categories,
and may not recognize the product names from the calls. Building a custom model to identify the
product keywords and then running them through a classification algorithm would require more
data preprocessing and development time, as well as more coding and testing. References:
* AutoML Natural Language documentation
* AI Platform Training built-in algorithms documentation
* Cloud Natural Language API documentation

Valid Professional-Machine-Learning-Engineer Dumps shared by ExamDiscuss.com for


Helping Passing Professional-Machine-Learning-Engineer Exam! ExamDiscuss.com now offer
the newest Professional-Machine-Learning-Engineer exam dumps, the ExamDiscuss.com
Professional-Machine-Learning-Engineer exam questions have been updated and answers
have been corrected get the newest ExamDiscuss.com Professional-Machine-Learning-
Engineer dumps with Test Engine here:
https://www.examdiscuss.com/Google/exam/Professional-Machine-Learning-
Engineer/premium/ (273 Q&As Dumps, 35%OFF Special Discount Code: freecram)

NEW QUESTION: 77
You recently developed a deep learning model using Keras, and now you are experimenting with
different training strategies. First, you trained the model using a single GPU, but the training
process was too slow.
Next, you distributed the training across 4 GPUs using tf.distribute.MirroredStrategy (with no other
changes), but you did not observe a decrease in training time. What should you do?
A. Distribute the dataset with tf.distribute.Strategy.experimental_distribute_dataset
B. Create a custom training loop.
C. Use a TPU with tf.distribute.TPUStrategy.
D. Increase the batch size.
Answer: (SHOW ANSWER)
* Option A is incorrect because distributing the dataset with
tf.distribute.Strategy.experimental_distribute_dataset is not the most effective way to decrease
the training time. This method allows you to distribute your dataset across multiple devices or
machines, by creating a tf.data.Dataset instance that can be iterated over in parallel1. However,
this option may not improve the training time significantly, as it does not change the amount of
data or computation that each device or machine has to process. Moreover, this option may
introduce additional overhead or complexity, as it requires you to handle the data sharding,
replication, and synchronization across the devices or machines1.
* Option B is incorrect because creating a custom training loop is not the easiest way to decrease
the training time. A custom training loop is a way to implement your own logic for training your
model, by using low-level TensorFlow APIs, such as tf.GradientTape, tf.Variable, or tf.function2. A
custom training loop may give you more flexibility and control over the training process, but it also
requires more effort and expertise, as you have to write and debug the code for each step of the
training loop, such as computing the gradients, applying the optimizer, or updating the metrics2.
Moreover, a custom training loop may not improve the training time significantly, as it does not
change the amount of data or computation that each device or machine has to process.
* Option C is incorrect because using a TPU with tf.distribute.TPUStrategy is not a valid way to
decrease the training time. A TPU (Tensor Processing Unit) is a custom hardware accelerator
designed for high-performance ML workloads3. A tf.distribute.TPUStrategy is a distribution
strategy that allows you to distribute your training across multiple TPUs, by creating a
tf.distribute.TPUStrategy instance that can be used with high-level TensorFlow APIs, such as
Keras4. However, this option is not feasible, as Vertex AI Training does not support TPUs as
accelerators for custom training jobs5. Moreover, this option may require significant code
changes, as TPUs have different requirements and limitations than GPUs.
* Option D is correct because increasing the batch size is the best way to decrease the training
time. The batch size is a hyperparameter that determines how many samples of data are
processed in each iteration of the training loop. Increasing the batch size may reduce the training
time, as it reduces the number of iterations needed to train the model, and it allows each device
or machine to process more data in parallel. Increasing the batch size is also easy to implement,
as it only requires changing a single
* hyperparameter. However, increasing the batch size may also affect the convergence and the
accuracy of the model, so it is important to find the optimal batch size that balances the trade-off
between the training time and the model performance.
References:
* tf.distribute.Strategy.experimental_distribute_dataset
* Custom training loop
* TPU overview
* tf.distribute.TPUStrategy
* Vertex AI Training accelerators
* [TPU programming model]
* [Batch size and learning rate]
* [Keras overview]
* [tf.distribute.MirroredStrategy]
* [Vertex AI Training overview]
* [TensorFlow overview]

NEW QUESTION: 78
You are working on a classification problem with time series data and achieved an area under the
receiver operating characteristic curve (AUC ROC) value of 99% for training data after just a few
experiments. You haven't explored using any sophisticated algorithms or spent any time on
hyperparameter tuning. What should your next step be to identify and fix the problem?
A. Address the model overfitting by using a less complex algorithm.
B. Address data leakage by applying nested cross-validation during model training.
C. Address data leakage by removing features highly correlated with the target value.
D. Address the model overfitting by tuning the hyperparameters to reduce the AUC ROC value.
Answer: (SHOW ANSWER)
Data leakage is a problem where information from outside the training dataset is used to create
the model, resulting in an overly optimistic or invalid estimate of the model performance. Data
leakage can occur in time series data when the temporal order of the data is not preserved during
data preparation or model evaluation.
For example, if the data is shuffled before splitting into train and test sets, or if future data is used
to impute missing values in past data, then data leakage can occur.
One way to address data leakage in time series data is to apply nested cross-validation during
model training.
Nested cross-validation is a technique that allows you to perform both model selection and model
evaluation in a robust way, while preserving the temporal order of the data. Nested cross-
validation involves two levels of cross-validation: an inner loop for model selection and an outer
loop for model evaluation. The inner loop splits the training data into k folds, trains and tunes the
model on k-1 folds, and validates the model on the remaining fold. The inner loop repeats this
process for each fold and selects the best model based on the validation performance. The outer
loop splits the data into n folds, trains the best model from the inner loop on n-1 folds, and tests
the model on the remaining fold. The outer loop repeats this process for each fold and evaluates
the model performance based on the test results.
Nested cross-validation can help to avoid data leakage in time series data by ensuring that the
model is trained and tested on non-overlapping data, and that the data used for validation is
never seen by the model during training. Nested cross-validation can also provide a more reliable
estimate of the model performance than a single train-test split or a simple cross-validation, as it
reduces the variance and bias of the estimate.
References:
* Data Leakage in Machine Learning
* How to Avoid Data Leakage When Performing Data Preparation
* Classification on a single time series - prevent leakage between train and test

NEW QUESTION: 79
You have been asked to develop an input pipeline for an ML training model that processes
images from disparate sources at a low latency. You discover that your input data does not fit in
memory. How should you create a dataset following Google-recommended best practices?
A. Create a tf.data.Dataset.prefetch transformation
B. Convert the images to tf .Tensor Objects, and then run Dataset. from_tensor_slices{).
C. Convert the images to tf .Tensor Objects, and then run tf. data. Dataset. from_tensors ().
D. Convert the images Into TFRecords, store the images in Cloud Storage, and then use the tf.
data API to read the images for training
Answer: (SHOW ANSWER)
An input pipeline is a way to prepare and feed data to a machine learning model for training or
inference. An input pipeline typically consists of several steps, such as reading, parsing,
transforming, batching, and prefetching the data. An input pipeline can improve the performance
and efficiency of the model, as it can handle large and complex datasets, optimize the data
processing, and reduce the latency and memory usage1.
For the use case of developing an input pipeline for an ML training model that processes images
from disparate sources at a low latency, the best option is to convert the images into TFRecords,
store the images in Cloud Storage, and then use the tf.data API to read the images for training.
This option involves using the following components and techniques:
* TFRecords: TFRecords is a binary file format that can store a sequence of data records, such
as images, text, or audio. TFRecords can help to compress, serialize, and store the data
efficiently, and reduce the data loading and parsing time. TFRecords can also support data
sharding and interleaving, which can improve the data throughput and parallelism2.
* Cloud Storage: Cloud Storage is a service that allows you to store and access data on Google
Cloud.
Cloud Storage can help to store and manage large and distributed datasets, such as images from
different sources, and provide high availability, durability, and scalability. Cloud Storage can also
integrate with other Google Cloud services, such as Compute Engine, AI Platform, and
Dataflow3.
* tf.data API: tf.data API is a set of tools and methods that allow you to create and manipulate
data pipelines in TensorFlow. tf.data API can help to read, transform, batch, and prefetch the data
efficiently, and optimize the data processing for performance and memory. tf.data API can also
support various data sources and formats, such as TFRecords, CSV, JSON, and images.
By using these components and techniques, the input pipeline can process large datasets of
images from disparate sources that do not fit in memory, and provide low latency and high
performance for the ML training model. Therefore, converting the images into TFRecords, storing
the images in Cloud Storage, and using the tf.data API to read the images for training is the best
option for this use case.
References:
* Build TensorFlow input pipelines | TensorFlow Core
* TFRecord and tf.Example | TensorFlow Core
* Cloud Storage documentation | Google Cloud
* [tf.data: Build TensorFlow input pipelines | TensorFlow Core]

NEW QUESTION: 80
A. Differential privacy
B. Federated learning
C. MD5 to encrypt data
D. Data Loss Prevention API
Answer: (SHOW ANSWER)
Federated learning is a machine learning technique that enables organizations to train AI models
on decentralized data without centralizing or sharing it1. It allows data privacy, continual learning,
and better performance on end-user devices2. Federated learning works by sending the model
parameters to the devices, where they are updated locally on the device's data, and then
aggregating the updated parameters on a central server to form a global model3. This way, the
data never leaves the device and the model can learn from a large and diverse dataset.
Federated learning is suitable for the use case of building an ML-based biometric authentication
for the bank's mobile app that verifies a customer's identity based on their fingerprint. Fingerprints
are considered highly sensitive personal information and cannot be downloaded and stored into
the bank databases. By using federated learning, the bank can train and deploy an ML model that
can recognize fingerprints without compromising the data privacy of the customers. The model
can also adapt to the variations and changes in the fingerprints over time and improve its
accuracy and reliability. Therefore, federated learning is the best learning strategy for this use
case.

NEW QUESTION: 81
You are developing ML models with Al Platform for image segmentation on CT scans. You
frequently update your model architectures based on the newest available research papers, and
have to rerun training on the same dataset to benchmark their performance. You want to
minimize computation costs and manual intervention while having version control for your code.
What should you do?
A. Use Cloud Functions to identify changes to your code in Cloud Storage and trigger a retraining
job
B. Use the gcloud command-line tool to submit training jobs on Al Platform when you update your
code
C. Use Cloud Build linked with Cloud Source Repositories to trigger retraining when new code is
pushed to the repository
D. Create an automated workflow in Cloud Composer that runs daily and looks for changes in
code in Cloud Storage using a sensor.
Answer: (SHOW ANSWER)
Developing ML models with AI Platform for image segmentation on CT scans requires a lot of
computation and experimentation, as image segmentation is a complex and challenging task that
involves assigning a label to each pixel in an image. Image segmentation can be used for various
medical applications, such as tumor detection, organ segmentation, or lesion localization1 To
minimize the computation costs and manual intervention while having version control for the
code, one should use Cloud Build linked with Cloud Source Repositories to trigger retraining
when new code is pushed to the repository. Cloud Build is a service that executes your builds on
Google Cloud Platform infrastructure. Cloud Build can import source code from Cloud Source
Repositories, Cloud Storage, GitHub, or Bitbucket, execute a build to your specifications, and
produce artifacts such as Docker containers or Java archives2 Cloud Build allows you to set up
automated triggers that start a build when changes are pushed to a source code repository. You
can configure triggers to filter the changes based on the branch, tag, or file path3 Cloud Source
Repositories is a service that provides fully managed private Git repositories on Google Cloud
Platform. Cloud Source Repositories allows you to store, manage, and track your code using the
Git version control system. You can also use Cloud Source Repositories to connect to other
Google Cloud services, such as Cloud Build, Cloud Functions, or Cloud Run4 To use Cloud Build
linked with Cloud Source Repositories to trigger retraining when new code is pushed to the
repository, you need to do the following steps:
* Create a Cloud Source Repository for your code, and push your code to the repository. You can
use the Cloud SDK, Cloud Console, or Cloud Source Repositories API to create and manage
your repository5
* Create a Cloud Build trigger for your repository, and specify the build configuration and the
trigger
* settings. You can use the Cloud SDK, Cloud Console, or Cloud Build API to create and manage
your trigger.
* Specify the steps of the build in a YAML or JSON file, such as installing the dependencies,
running the tests, building the container image, and submitting the training job to AI Platform. You
can also use the Cloud Build predefined or custom build steps to simplify your build configuration.
* Push your new code to the repository, and the trigger will start the build automatically. You can
monitor the status and logs of the build using the Cloud SDK, Cloud Console, or Cloud Build API.
The other options are not as easy or feasible. Using Cloud Functions to identify changes to your
code in Cloud Storage and trigger a retraining job is not ideal, as Cloud Functions has limitations
on the memory, CPU, and execution time, and does not provide a user interface for managing
and tracking your builds. Using the gcloud command-line tool to submit training jobs on AI
Platform when you update your code is not optimal, as it requires manual intervention and does
not leverage the benefits of Cloud Build and its integration with Cloud Source Repositories.
Creating an automated workflow in Cloud Composer that runs daily and looks for changes in code
in Cloud Storage using a sensor is not relevant, as Cloud Composer is mainly designed for
orchestrating complex workflows across multiple systems, and does not provide a version control
system for your code.
References: 1: Image segmentation 2: Cloud Build overview 3: Creating and managing build
triggers 4: Cloud Source Repositories overview 5: Quickstart: Create a repository : [Quickstart:
Create a build trigger] :
[Configuring builds] : [Viewing build results]

NEW QUESTION: 82
You need to build an ML model for a social media application to predict whether a user's
submitted profile photo meets the requirements. The application will inform the user if the picture
meets the requirements. How should you build a model to ensure that the application does not
falsely accept a non-compliant picture?
A. Use AutoML to optimize the model's recall in order to minimize false negatives.
B. Use AutoML to optimize the model's F1 score in order to balance the accuracy of false
positives and false negatives.
C. Use Vertex AI Workbench user-managed notebooks to build a custom model that has three
times as many examples of pictures that meet the profile photo requirements.
D. Use Vertex AI Workbench user-managed notebooks to build a custom model that has three
times as many examples of pictures that do not meet the profile photo requirements.
Answer: (SHOW ANSWER)
Recall is the ratio of true positives to the sum of true positives and false negatives. It measures
how well the model can identify all the relevant cases. In this scenario, the relevant cases are the
pictures that do not meet the profile photo requirements. Therefore, minimizing false negatives
means minimizing the cases where the model incorrectly predicts that a non-compliant picture
meets the requirements. By using AutoML to optimize the model's recall, the model will be more
likely to reject a non-compliant picture and inform the user accordingly. References:
* [AutoML Vision] is a service that allows you to train custom ML models for image classification
and object detection tasks. You can use AutoML to optimize your model for different metrics, such
as recall, precision, or F1 score.
* [Recall] is one of the evaluation metrics for ML models. It is defined as TP / (TP + FN), where
TP is the number of true positives and FN is the number of false negatives. Recall measures how
well the model can identify all the relevant cases. A high recall means that the model has a low
rate of false negatives.

NEW QUESTION: 83
You work for a biotech startup that is experimenting with deep learning ML models based on
properties of biological organisms. Your team frequently works on early-stage experiments with
new architectures of ML models, and writes custom TensorFlow ops in C++. You train your
models on large datasets and large batch sizes. Your typical batch size has 1024 examples, and
each example is about 1 MB in size. The average size of a network with all weights and
embeddings is 20 GB. What hardware should you choose for your models?
A. A cluster with 2 n1-highcpu-64 machines, each with 8 NVIDIA Tesla V100 GPUs (128 GB GPU
memory in total), and a n1-highcpu-64 machine with 64 vCPUs and 58 GB RAM
B. A cluster with 2 a2-megagpu-16g machines, each with 16 NVIDIA Tesla A100 GPUs (640 GB
GPU memory in total), 96 vCPUs, and 1.4 TB RAM
C. A cluster with an n1-highcpu-64 machine with a v2-8 TPU and 64 GB RAM
D. A cluster with 4 n1-highcpu-96 machines, each with 96 vCPUs and 86 GB RAM
Answer: (SHOW ANSWER)
The best hardware to choose for your models is a cluster with 2 a2-megagpu-16g machines, each
with 16 NVIDIA Tesla A100 GPUs (640 GB GPU memory in total), 96 vCPUs, and 1.4 TB RAM.
This hardware configuration can provide you with enough compute power, memory, and
bandwidth to handle your large and complex deep learning models, as well as your custom
TensorFlow ops in C++. The NVIDIA Tesla A100 GPUs are the latest and most advanced GPUs
from NVIDIA, which offer high performance, scalability, and efficiency for various ML workloads.
They also support multi-instance GPU (MIG) technology, which allows you to partition each GPU
into up to seven smaller instances, each with its own memory, cache, and compute cores. This
can enable you to run multiple experiments in parallel, or to optimize the resource utilization and
cost efficiency of your models. The a2-megagpu-16g machines are part of the Google Cloud
Accelerator-Optimized VM (A2) family, which are designed to provide the best performance and
flexibility for GPU-intensive applications. They also offer high-speed NVLink interconnects
between the GPUs, which can improve the data transfer and communication between the GPUs.
Moreover, the a2-megagpu-16g machines have 96 vCPUs and 1.4 TB RAM, which can support
the CPU and memory requirements of your models, as well as the data preprocessing and
postprocessing tasks.
The other options are not optimal for the following reasons:
* A. A cluster with 2 n1-highcpu-64 machines, each with 8 NVIDIA Tesla V100 GPUs (128 GB
GPU memory in total), and a n1-highcpu-64 machine with 64 vCPUs and 58 GB RAM is not a
good option, as it has less GPU memory, compute power, and bandwidth than the a2-
megagpu-16g machines. The NVIDIA Tesla V100 GPUs are the previous generation of GPUs
from NVIDIA, which have lower performance, scalability, and efficiency than the NVIDIA Tesla
A100 GPUs. They also do not support the MIG technology, which can limit the flexibility and
optimization of your models. Moreover, the n1-highcpu-64 machines are part of the Google Cloud
N1 VM family, which are general-purpose VMs that do not offer the best performance and
features for GPU-intensive applications. They also have lower vCPUs and RAM than the a2-
megagpu-16g machines, which can affect the CPU and memory requirements of your models, as
well as the data preprocessing and postprocessing tasks.
* C. A cluster with an n1-highcpu-64 machine with a v2-8 TPU and 64 GB RAM is not a good
option, as it has less GPU memory, compute power, and bandwidth than the a2-megagpu-16g
machines. The v2-8
* TPU is a cloud tensor processing unit (TPU) device, which is a custom ASIC chip designed by
Google to accelerate ML workloads. However, the v2-8 TPU is the second generation of TPUs,
which have lower performance, scalability, and efficiency than the latest v3-8 TPUs. They also
have less memory and bandwidth than the NVIDIA Tesla A100 GPUs, which can limit the size
and complexity of your models, as well as the data transfer and communication between the
devices. Moreover, the n1-highcpu-64 machine has lower vCPUs and RAM than the a2-
megagpu-16g machines, which can affect the CPU and memory requirements of your models, as
well as the data preprocessing and postprocessing tasks.
* D. A cluster with 4 n1-highcpu-96 machines, each with 96 vCPUs and 86 GB RAM is not a good
option, as it does not have any GPUs, which are essential for accelerating deep learning models.
The n1-highcpu-96 machines are part of the Google Cloud N1 VM family, which are general-
purpose VMs that do not offer the best performance and features for GPU-intensive applications.
They also have lower RAM than the a2-megagpu-16g machines, which can affect the memory
requirements of your models, as well as the data preprocessing and postprocessing tasks.
References:
* Professional ML Engineer Exam Guide
* Preparing for Google Cloud Certification: Machine Learning Engineer Professional Certificate
* Google Cloud launches machine learning engineer certification
* NVIDIA Tesla A100 GPU
* Google Cloud Accelerator-Optimized VM (A2) family
* Google Cloud N1 VM family
* Cloud TPU

NEW QUESTION: 84
You have created a Vertex Al pipeline that automates custom model training You want to add a
pipeline component that enables your team to most easily collaborate when running different
executions and comparing metrics both visually and programmatically. What should you do?
A. Add a component to the Vertex Al pipeline that logs metrics to a BigQuery table Query the
table to compare different executions of the pipeline Connect BigQuery to Looker Studio to
visualize metrics.
B. Add a component to the Vertex Al pipeline that logs metrics to a BigQuery table Load the table
into a pandas DataFrame to compare different executions of the pipeline Use Matplotlib to
visualize metrics.
C. Add a component to the Vertex Al pipeline that logs metrics to Vertex ML Metadata Use Vertex
Al Experiments to compare different executions of the pipeline Use Vertex Al TensorBoard to
visualize metrics.
D. Add a component to the Vertex Al pipeline that logs metrics to Vertex ML Metadata Load the
Vertex ML Metadata into a pandas DataFrame to compare different executions of the pipeline.
Use Matplotlib to visualize metrics.
Answer: (SHOW ANSWER)
Vertex AI Experiments is a managed service that allows you to track, compare, and manage
experiments with Vertex AI. You can use Vertex AI Experiments to record the parameters,
metrics, and artifacts of each pipeline run, and compare them in a graphical interface. Vertex AI
TensorBoard is a tool that lets you visualize the metrics of your models, such as accuracy, loss,
and learning curves. By logging metrics to Vertex ML Metadata and using Vertex AI Experiments
and TensorBoard, you can easily collaborate with your team and find the best model configuration
for your problem. References: Vertex AI Pipelines: Metrics visualization and run comparison using
the KFP SDK, Track, compare, manage experiments with Vertex AI Experiments, Vertex AI
Pipelines

NEW QUESTION: 85
You recently deployed a model lo a Vertex Al endpoint and set up online serving in Vertex Al
Feature Store.
You have configured a daily batch ingestion job to update your featurestore During the batch
ingestion jobs you discover that CPU utilization is high in your featurestores online serving nodes
and that feature retrieval latency is high. You need to improve online serving performance during
the daily batch ingestion. What should you do?
A. Schedule an increase in the number of online serving nodes in your featurestore prior to the
batch ingestion jobs.
B. Enable autoscaling of the online serving nodes in your featurestore
C. Enable autoscaling for the prediction nodes of your DeployedModel in the Vertex Al endpoint.
D. Increase the worker counts in the importFeaturevalues request of your batch ingestion job.
Answer: (SHOW ANSWER)
Vertex AI Feature Store provides two options for online serving: Bigtable and optimized online
serving. Both options support autoscaling, which means that the number of online serving nodes
can automatically adjust to the traffic demand. By enabling autoscaling, you can improve the
online serving performance and reduce the feature retrieval latency during the daily batch
ingestion. Autoscaling also helps you optimize the cost and resource utilization of your
featurestore. References:
* Online serving | Vertex AI | Google Cloud
* New Vertex AI Feature Store: BigQuery-Powered, GenAI-Ready | Google Cloud Blog

NEW QUESTION: 86
You are building a TensorFlow model for a financial institution that predicts the impact of
consumer spending on inflation globally. Due to the size and nature of the data, your model is
long-running across all types of hardware, and you have built frequent checkpointing into the
training process. Your organization has asked you to minimize cost. What hardware should you
choose?
A. A Vertex AI Workbench user-managed notebooks instance running on an n1-standard-16 with
4 NVIDIA P100 GPUs
B. A Vertex AI Workbench user-managed notebooks instance running on an n1-standard-16 with
an NVIDIA P100 GPU
C. A Vertex AI Workbench user-managed notebooks instance running on an n1-standard-16 with
a non-preemptible v3-8 TPU
D. A Vertex AI Workbench user-managed notebooks instance running on an n1-standard-16 with
a preemptible v3-8 TPU
Answer: (SHOW ANSWER)
The best hardware to choose for your model while minimizing cost is a Vertex AI Workbench
user-managed notebooks instance running on an n1-standard-16 with a preemptible v3-8 TPU.
This hardware configuration can provide you with high performance, scalability, and efficiency for
your TensorFlow model, as well as low cost and flexibility for your long-running and checkpointing
process. The v3-8 TPU is a cloud tensor processing unit (TPU) device, which is a custom ASIC
chip designed by Google to accelerate ML workloads.
It can handle large and complex models and datasets, and offer fast and stable training and
inference. The n1-standard-16 is a general-purpose VM that can support the CPU and memory
requirements of your model, as well as the data preprocessing and postprocessing tasks. By
choosing a preemptible v3-8 TPU, you can take advantage of the lower price and availability of
the TPU devices, as long as you can tolerate the possibility of the device being reclaimed by
Google at any time. However, since you have built frequent checkpointing into your training
process, you can resume your model from the last saved state, and avoid losing any progress or
data. Moreover, you can use the Vertex AI Workbench user-managed notebooks to create and
manage your notebooks instances, and leverage the integration with Vertex AI and other Google
Cloud services.
The other options are not optimal for the following reasons:
* A. A Vertex AI Workbench user-managed notebooks instance running on an n1-standard-16
with 4 NVIDIA P100 GPUs is not a good option, as it has higher cost and lower performance than
the v3-8 TPU. The NVIDIA P100 GPUs are the previous generation of GPUs from NVIDIA, which
have lower performance, scalability, and efficiency than the latest NVIDIA A100 GPUs or the
TPUs. They also have higher price and lower availability than the preemptible TPUs, which can
increase the cost and complexity of your solution.
* B. A Vertex AI Workbench user-managed notebooks instance running on an n1-standard-16
with an NVIDIA P100 GPU is not a good option, as it has higher cost and lower performance than
the v3-8 TPU. It also has less GPU memory and compute power than the option with 4 NVIDIA
P100 GPUs, which can limit the size and complexity of your model, and affect the training and
inference speed and quality.
* C. A Vertex AI Workbench user-managed notebooks instance running on an n1-standard-16
with a
* non-preemptible v3-8 TPU is not a good option, as it has higher cost and lower flexibility than
the preemptible v3-8 TPU. The non-preemptible v3-8 TPU has the same performance, scalability,
and efficiency as the preemptible v3-8 TPU, but it has higher price and lower availability, as it is
reserved for your exclusive use. Moreover, since your model is long-running and checkpointing,
you do not need the guarantee of the device not being reclaimed by Google, and you can benefit
from the lower cost and higher availability of the preemptible v3-8 TPU.
References:
* Professional ML Engineer Exam Guide
* Preparing for Google Cloud Certification: Machine Learning Engineer Professional Certificate
* Google Cloud launches machine learning engineer certification
* Cloud TPU
* Vertex AI Workbench user-managed notebooks
* Preemptible VMs
* NVIDIA Tesla P100 GPU

NEW QUESTION: 87
You work with a team of researchers to develop state-of-the-art algorithms for financial analysis.
Your team develops and debugs complex models in TensorFlow. You want to maintain the ease
of debugging while also reducing the model training time. How should you set up your training
environment?
A. Configure a v3-8 TPU VM SSH into the VM to tram and debug the model.
B. Configure a v3-8 TPU node Use Cloud Shell to SSH into the Host VM to train and debug the
model.
C. Configure a M-standard-4 VM with 4 NVIDIA P100 GPUs SSH into the VM and use Parameter
Server Strategy to train the model.
D. Configure a M-standard-4 VM with 4 NVIDIA P100 GPUs SSH into the VM and use
MultiWorkerMirroredStrategy to train the model.
Answer: (SHOW ANSWER)
A TPU VM is a virtual machine that has direct access to a Cloud TPU device. TPU VMs provide a
simpler and more flexible way to use Cloud TPUs, as they eliminate the need for a separate host
VM and network setup. TPU VMs also support interactive debugging tools such as TensorFlow
Debugger (tfdbg) and Python Debugger (pdb), which can help researchers develop and
troubleshoot complex models. A v3-8 TPU VM has
8 TPU cores, which can provide high performance and scalability for training large models.
SSHing into the TPU VM allows the user to run and debug the TensorFlow code directly on the
TPU device, without any network overhead or data transfer issues. References:
* 1: TPU VMs Overview
* 2: TPU VMs Quickstart
* 3: Debugging TensorFlow Models on Cloud TPUs

NEW QUESTION: 88
Your work for a textile manufacturing company. Your company has hundreds of machines and
each machine has many sensors. Your team used the sensory data to build hundreds of ML
models that detect machine anomalies Models are retrained daily and you need to deploy these
models in a cost-effective way. The models must operate 24/7 without downtime and make sub
millisecond predictions. What should you do?
A. Deploy a Dataflow batch pipeline and a Vertex Al Prediction endpoint.
B. Deploy a Dataflow batch pipeline with the Runlnference API. and use model refresh.
C. Deploy a Dataflow streaming pipeline and a Vertex Al Prediction endpoint with autoscaling.
D. Deploy a Dataflow streaming pipeline with the Runlnference API and use automatic model
refresh.
Answer: (SHOW ANSWER)
A Dataflow streaming pipeline is a cost-effective way to process large volumes of real-time data
from sensors.
The RunInference API is a Dataflow transform that allows you to run online predictions on your
streaming data using your ML models. By using the RunInference API, you can avoid the latency
and cost of using a separate prediction service. The automatic model refresh feature enables you
to update your models in the pipeline without redeploying the pipeline. This way, you can ensure
that your models are always up-to-date and accurate. By deploying a Dataflow streaming pipeline
with the RunInference API and using automatic model refresh, you can achieve sub-millisecond
predictions, 24/7 availability, and low operational overhead for your ML models. References:
* Dataflow documentation
* RunInference API documentation
* Automatic model refresh documentation
* Preparing for Google Cloud Certification: Machine Learning Engineer Professional Certificate

NEW QUESTION: 89
You work as an analyst at a large banking firm. You are developing a robust, scalable ML pipeline
to train several regression and classification models. Your primary focus for the pipeline is model
interpretability. You want to productionize the pipeline as quickly as possible What should you
do?
A. Use Tabular Workflow for Wide & Deep through Vertex Al Pipelines to jointly train wide linear
models and deep neural networks.
B. Use Google Kubernetes Engine to build a custom training pipeline for XGBoost-based models.
C. Use Tabular Workflow forTabel through Vertex Al Pipelines to train attention-based models.
D. Use Cloud Composer to build the training pipelines for custom deep learning-based models.
Answer: (SHOW ANSWER)
According to the official exam guide1, one of the skills assessed in the exam is to "automate and
orchestrate ML pipelines using Cloud Composer". Cloud Composer2 is a fully managed workflow
orchestration service that uses Apache Airflow to create, schedule, monitor, and manage
workflows. Cloud Composer allows you to build custom training pipelines for deep learning-based
models and integrate them with other Google Cloud services. You can also use Cloud Composer
to implement model interpretability techniques, such as feature attributions, explainable AI, or
model debugging3. The other options are not relevant or optimal for this scenario. References:
* Professional ML Engineer Exam Guide
* Cloud Composer
* Model interpretability with Cloud Composer
* Google Professional Machine Learning Certification Exam 2023
* Latest Google Professional Machine Learning Engineer Actual Free Exam Questions

NEW QUESTION: 90
A. Run one hypertuning job for 100 trials Set num_hidden_layers and learning_rate as conditional
hyperparameters based on their parent hyperparameter training method.
B. Run one hypertuning job with training_method as the hyperparameter for 50 trials Select the
architecture with the lowest training loss. and further hypertune It and its corresponding
hyperparameters for 50 trials
C. Run two separate hypertuning jobs. a linear regression job for 50 trials, and a DNN job for 50
trials Compare their final performance on a common validation set. and select the set of
hyperparameters with the least training loss
D. Run one hypertuning job for 100 trials. Set num hidden_layers as a conditional hypetparameter
based on its parent hyperparameter training_mothod. and set learning rate as a non-conditional
hyperparameter
Answer: (SHOW ANSWER)

NEW QUESTION: 91
While performing exploratory data analysis on a dataset, you find that an important categorical
feature has 5% null values. You want to minimize the bias that could result from the missing
values. How should you handle the missing values?
A. Remove the rows with missing values, and upsample your dataset by 5%.
B. Replace the missing values with the feature's mean.
C. Replace the missing values with a placeholder category indicating a missing value.
D. Move the rows with missing values to your validation dataset.
Answer: (SHOW ANSWER)
The best option for handling missing values in a categorical feature is to replace them with a
placeholder category indicating a missing value. This is a type of imputation, which is a method of
estimating the missing values based on the observed data. Imputing the missing values with a
placeholder category preserves the information that the data is missing, and avoids introducing
bias or distortion in the feature distribution. It also allows the machine learning model to learn
from the missingness pattern, and potentially use it as a predictor for the target variable. The
other options are not suitable for handling missing values in a categorical feature, because:
* Removing the rows with missing values and upsampling the dataset by 5% would reduce the
size of the dataset and potentially lose important information. It would also introduce sampling
bias and overfitting, as the upsampling process would create duplicate or synthetic observations
that do not reflect the true population.
* Replacing the missing values with the feature's mean would not make sense for a categorical
feature, as
* the mean is a numerical measure that does not capture the mode or frequency of the
categories. It would also create a new category that does not exist in the original data, and might
confuse the machine learning model.
* Moving the rows with missing values to the validation dataset would compromise the validity and
reliability of the model evaluation, as the validation dataset would not be representative of the test
or production data. It would also reduce the amount of data available for training the model, and
might introduce leakage or inconsistency between the training and validation datasets.
References:
* Imputation of missing values
* Effective Strategies to Handle Missing Values in Data Analysis
* How to Handle Missing Values of Categorical Variables?
* Google Cloud launches machine learning engineer certification
* Google Professional Machine Learning Engineer Certification
* Professional ML Engineer Exam Guide
* Preparing for Google Cloud Certification: Machine Learning Engineer Professional Certificate

Valid Professional-Machine-Learning-Engineer Dumps shared by ExamDiscuss.com for


Helping Passing Professional-Machine-Learning-Engineer Exam! ExamDiscuss.com now offer
the newest Professional-Machine-Learning-Engineer exam dumps, the ExamDiscuss.com
Professional-Machine-Learning-Engineer exam questions have been updated and answers
have been corrected get the newest ExamDiscuss.com Professional-Machine-Learning-
Engineer dumps with Test Engine here:
https://www.examdiscuss.com/Google/exam/Professional-Machine-Learning-
Engineer/premium/ (273 Q&As Dumps, 35%OFF Special Discount Code: freecram)

NEW QUESTION: 92
You are building a real-time prediction engine that streams files which may contain Personally
Identifiable Information (Pll) to Google Cloud. You want to use the Cloud Data Loss Prevention
(DLP) API to scan the files. How should you ensure that the Pll is not accessible by unauthorized
individuals?
A. Stream all files to Google CloudT and then write the data to BigQuery Periodically conduct a
bulk scan of the table using the DLP API.
B. Stream all files to Google Cloud, and write batches of the data to BigQuery While the data is
being written to BigQuery conduct a bulk scan of the data using the DLP API.
C. Create two buckets of data Sensitive and Non-sensitive Write all data to the Non-sensitive
bucket Periodically conduct a bulk scan of that bucket using the DLP API, and move the sensitive
data to the Sensitive bucket
D. Create three buckets of data: Quarantine, Sensitive, and Non-sensitive Write all data to the
Quarantine bucket.
E. Periodically conduct a bulk scan of that bucket using the DLP API, and move the data to either
the Sensitive or Non-Sensitive bucket
Answer: (SHOW ANSWER)
The Cloud DLP API is a service that allows users to inspect, classify, and de-identify sensitive
data. It can be used to scan data in Cloud Storage, BigQuery, Cloud Datastore, and Cloud
Pub/Sub. The best way to ensure that the PII is not accessible by unauthorized individuals is to
use a quarantine bucket to store the data before scanning it with the DLP API. This way, the data
is isolated from other applications and users until it is classified and moved to the appropriate
bucket. The other options are not as secure or efficient, as they either expose the data to
BigQuery before scanning, or scan the data after writing it to a non-sensitive bucket.
References:
* Cloud DLP documentation
* Scanning and classifying Cloud Storage files

NEW QUESTION: 93
You are training a custom language model for your company using a large dataset. You plan to
use the ReductionServer strategy on Vertex Al. You need to configure the worker pools of the
distributed training job.
What should you do?
A. Configure the machines of the first two worker pools to have GPUs and to use a container
image where your training code runs Configure the third worker pool to have GPUs: and use the
reduction server container image.
B. Configure the machines of the first two worker pools to have GPUs and to use a container
image where your training code runs. Configure the third worker pool to use the reductionserver
container image without accelerators, and choose a machine type that prioritizes bandwidth.
C. Configure the machines of the first two worker pools to have TPUs and to use a container
image where your training code runs Configure the third worker pool without accelerators, and
use the reductionserver container image without accelerators and choose a machine type that
prioritizes bandwidth.
D. Configure the machines of the first two pools to have TPUs. and to use a container image
where your training code runs Configure the third pool to have TPUs: and use the reductionserver
container image.
Answer: B (LEAVE A REPLY)
According to the web search results, Reduction Server is a faster GPU all-reduce algorithm
developed at Google that uses a dedicated set of reducers to aggregate gradients from
workers12. Reducers are lightweight CPU VM instances that are significantly cheaper than GPU
VMs2. Therefore, the third worker pool should not have any accelerators, and should use a
machine type that has high network bandwidth to optimize the communication between workers
and reducers2. TPUs are not supported by Reduction Server, so the first two worker pools should
have GPUs and use a container image that contains the training code12. The reduction-server
container image is provided by Google and should be used for the third worker pool2.
NEW QUESTION: 94
You are an ML engineer at a global shoe store. You manage the ML models for the company's
website. You are asked to build a model that will recommend new products to the user based on
their purchase behavior and similarity with other users. What should you do?
A. Build a classification model
B. Build a knowledge-based filtering model
C. Build a collaborative-based filtering model
D. Build a regression model using the features as predictors
Answer: (SHOW ANSWER)
A recommender system is a type of machine learning system that suggests relevant items to
users based on their preferences and behavior. Recommender systems are widely used in e-
commerce, media, and entertainment industries to enhance user experience and increase
revenue1 There are different types of recommender systems that use different filtering methods
to generate recommendations. The most common types are:
* Content-based filtering: This method uses the features of the items and the users to find the
similarity between them. For example, a content-based recommender system for movies may use
the genre, director, cast, and ratings of the movies, and the preferences, demographics, and
history of the users, to recommend movies that are similar to the ones the user liked before2
* Collaborative filtering: This method uses the feedback and ratings of the users to find the
similarity between them and the items. For example, a collaborative filtering recommender system
for books may use the ratings of the users for different books, and recommend books that are
liked by other users who have similar ratings to the target user3
* Hybrid method: This method combines content-based and collaborative filtering methods to
overcome the limitations of each method and improve the accuracy and diversity of the
recommendations. For example, a hybrid recommender system for music may use both the
features of the songs and the artists, and the ratings and listening habits of the users, to
recommend songs that match the user's taste and preferences4
* Deep learning-based: This method uses deep neural networks to learn complex and non-linear
patterns from the data and generate recommendations. Deep learning-based recommender
systems can handle large-scale and high-dimensional data, and incorporate various types of
information, such as text, images, audio, and video. For example, a deep learning-based
recommender system for fashion may use the images and descriptions of the products, and the
profiles and feedback of the users, to recommend products that suit the user's style and
preferences.
For the use case of building a model that will recommend new products to the user based on their
purchase behavior and similarity with other users, the best option is to build a collaborative-based
filtering model. This is because collaborative filtering can leverage the implicit feedback and
ratings of the users to find the items that are most likely to interest them. Collaborative filtering
can also help discover new products that the user may not be aware of, and increase the diversity
and serendipity of the recommendations3 The other options are not as suitable for this use case.
Building a classification model or a regression model using the features as predictors is not a
good idea, as these models are not designed for recommendation tasks, and may not capture the
preferences and behavior of the users. Building a knowledge-based filtering model is not relevant,
as this method uses the explicit knowledge and requirements of the users to find the items that
meet their criteria, and does not rely on the purchase behavior or similarity with other users.
References: 1: Recommender system 2: Content-based filtering 3: Collaborative filtering 4: Hybrid
recommender system : [Deep learning for recommender systems] : [Knowledge-based
recommender system]

NEW QUESTION: 95
You work on a team that builds state-of-the-art deep learning models by using the TensorFlow
framework.
Your team runs multiple ML experiments each week which makes it difficult to track the
experiment runs.
You want a simple approach to effectively track, visualize and debug ML experiment runs on
Google Cloud while minimizing any overhead code. How should you proceed?
A. Set up Vertex Al Experiments to track metrics and parameters Configure Vertex Al
TensorBoard for visualization.
B. Set up a Cloud Function to write and save metrics files to a Cloud Storage Bucket Configure a
Google Cloud VM to host TensorBoard locally for visualization.
C. Set up a Vertex Al Workbench notebook instance Use the instance to save metrics data in a
Cloud Storage bucket and to host TensorBoard locally for visualization.
D. Set up a Cloud Function to write and save metrics files to a BigQuery table. Configure a
Google Cloud VM to host TensorBoard locally for visualization.
Answer: (SHOW ANSWER)
Vertex AI Experiments is a service that allows you to track, compare, and optimize your ML
experiments on Google Cloud. You can use Vertex AI Experiments to log metrics and parameters
from your TensorFlow models, and then visualize them in Vertex AI TensorBoard. Vertex AI
TensorBoard is a managed service that provides a web interface for viewing and debugging your
ML experiments. You can use Vertex AI TensorBoard to compare different runs, inspect model
graphs, analyze scalars, histograms, images, and more.
By using Vertex AI Experiments and Vertex AI TensorBoard, you can simplify your ML experiment
tracking and visualization workflow, and avoid the overhead of setting up and maintaining your
own Cloud Functions, Cloud Storage buckets, or VMs. References:
* [Vertex AI Experiments documentation]
* [Vertex AI TensorBoard documentation]
* Preparing for Google Cloud Certification: Machine Learning Engineer Professional Certificate

NEW QUESTION: 96
You have recently used TensorFlow to train a classification model on tabular data You have
created a Dataflow pipeline that can transform several terabytes of data into training or prediction
datasets consisting of TFRecords. You now need to productionize the model, and you want the
predictions to be automatically uploaded to a BigQuery table on a weekly schedule. What should
you do?
A. Import the model into Vertex Al and deploy it to a Vertex Al endpoint On Vertex Al Pipelines
create a pipeline that uses the Dataf lowPythonJobop and the Mcdei3archPredictoc components.
B. Import the model into Vertex Al and deploy it to a Vertex Al endpoint Create a Dataflow
pipeline that reuses the data processing logic sends requests to the endpoint and then uploads
predictions to a BigQuery table.
C. Import the model into Vertex Al On Vertex Al Pipelines, create a pipeline that uses the
DatafIowPythonJobOp and the ModelBatchPredictOp components.
D. Import the model into BigQuery Implement the data processing logic in a SQL query On Vertex
Al Pipelines create a pipeline that uses the BigqueryQueryJobop and the
EigqueryPredictModejobOp components.
Answer: (SHOW ANSWER)
Vertex AI is a service that allows you to create and train ML models using Google Cloud
technologies. You can use Vertex AI to import the model that you trained with TensorFlow and
store it in the Vertex AI Model Registry. The Vertex AI Model Registry is a service that allows you
to store and manage your ML models on Google Cloud. You can then use Vertex AI Pipelines to
create a pipeline that uses the DataflowPythonJobOp and the ModelBatchPredictOp components.
The DataflowPythonJobOp component is a component that allows you to run a Dataflow job using
a Python script. Dataflow is a service that allows you to create and run scalable and portable data
processing pipelines on Google Cloud. You can use the DataflowPythonJobOp component to
reuse the data processing logic that you created for transforming the data into TFRecords. The
ModelBatchPredictOp component is a component that allows you to run a batch prediction job
using a model from the Vertex AI Model Registry. Batch prediction is a type of prediction that
provides high-throughput responses to large batches of input data. You can use the
ModelBatchPredictOp component to make predictions using the TFRecords from the
DataflowPythonJobOp component and the model from the Vertex AI Model Registry. You can
also configure the ModelBatchPredictOp component to automatically upload the predictions to a
BigQuery table. BigQuery is a service that allows you to store and query large amounts of data in
a scalable and cost-effective way. You can use BigQuery to store and analyze the predictions
from your model. You can also schedule the pipeline to run on a weekly basis, so that the
predictions are updated regularly. By using Vertex AI, Vertex AI Pipelines, Dataflow, and
BigQuery, you can productionize the model and upload the predictions to a BigQuery table on a
weekly schedule. References:
* Vertex AI documentation
* Vertex AI Pipelines documentation
* Dataflow documentation
* BigQuery documentation
* Preparing for Google Cloud Certification: Machine Learning Engineer Professional Certificate
NEW QUESTION: 97
You created an ML pipeline with multiple input parameters. You want to investigate the tradeoffs
between different parameter combinations. The parameter options are
* input dataset
* Max tree depth of the boosted tree regressor
* Optimizer learning rate
You need to compare the pipeline performance of the different parameter combinations measured
in F1 score, time to train and model complexity. You want your approach to be reproducible and
track all pipeline runs on the same platform. What should you do?
A. 1 Use BigQueryML to create a boosted tree regressor and use the hyperparameter tuning
capability
2 Configure the hyperparameter syntax to select different input datasets. max tree depths, and
optimizer teaming rates Choose the grid search option
B. 1 Create a Vertex Al pipeline with a custom model training job as part of the pipeline Configure
the pipeline's parameters to include those you are investigating
2 In the custom training step, use the Bayesian optimization method with F1 score as the target to
maximize
C. 1 Create a Vertex Al Workbench notebook for each of the different input datasets
2 In each notebook, run different local training jobs with different combinations of the max tree
depth and optimizer learning rate parameters
3 After each notebook finishes, append the results to a BigQuery table
D. 1 Create an experiment in Vertex Al Experiments
2. Create a Vertex Al pipeline with a custom model training job as part of the pipeline. Configure
the pipelines parameters to include those you are investigating
3. Submit multiple runs to the same experiment using different values for the parameters
Answer: (SHOW ANSWER)
The best option for investigating the tradeoffs between different parameter combinations is to
create an experiment in Vertex AI Experiments, create a Vertex AI pipeline with a custom model
training job as part of the pipeline, configure the pipeline's parameters to include those you are
investigating, and submit multiple runs to the same experiment using different values for the
parameters. This option allows you to leverage the power and flexibility of Google Cloud to
compare the pipeline performance of the different parameter combinations measured in F1 score,
time to train, and model complexity. Vertex AI Experiments is a service that can track and
compare the results of multiple machine learning runs. Vertex AI Experiments can record the
metrics, parameters, and artifacts of each run, and display them in a dashboard for easy
visualization and analysis. Vertex AI Experiments can also help users optimize the
hyperparameters of their models by using different search algorithms, such as grid search,
random search, or Bayesian optimization1. Vertex AI Pipelines is a service that can orchestrate
machine learning workflows using Vertex AI. Vertex AI Pipelines can run preprocessing and
training steps on custom Docker images, and evaluate, deploy, and monitor the machine learning
model. A custom model training job is a type of pipeline step that can train a custom model by
using a user-provided script or container. A custom model training job can accept pipeline
parameters as inputs, which can be used to control the training logic or data source. By creating
an experiment in Vertex AI Experiments, creating a Vertex AI pipeline with a custom model
training job as part of the pipeline, configuring the pipeline's parameters to include those you are
investigating, and submitting multiple runs to the same experiment using different values for the
parameters, you can create a reproducible and trackable approach to investigate the tradeoffs
between different parameter combinations.
The other options are not as good as option D, for the following reasons:
* Option A: Using BigQuery ML to create a boosted tree regressor and use the hyperparameter
tuning capability, configuring the hyperparameter syntax to select different input datasets, max
tree depths, and optimizer learning rates, and choosing the grid search option would not be able
to handle different input datasets as a hyperparameter, and would not be as flexible and scalable
as using Vertex AI Experiments and Vertex AI Pipelines. BigQuery ML is a service that can create
and train machine learning models by using SQL queries on BigQuery. BigQuery ML can perform
hyperparameter tuning by using the ML.FORECAST or ML.PREDICT functions, and specifying
the hyperparameters option. BigQuery ML can also use different search algorithms, such as grid
search, random search, or Bayesian optimization, to find the optimal hyperparameters. However,
BigQuery ML can only tune the hyperparameters that are related to the model architecture or
training process, such as max tree depth or learning rate. BigQuery ML cannot tune the
hyperparameters that are related to the data source, such as input dataset. Moreover, BigQuery
ML is not designed to work with Vertex AI Experiments or Vertex AI Pipelines, which can provide
more features and flexibility for tracking and orchestrating machine learning workflows2.
* Option B: Creating a Vertex AI pipeline with a custom model training job as part of the pipeline,
configuring the pipeline's parameters to include those you are investigating, and using the
Bayesian optimization method with F1 score as the target to maximize in the custom training step
would not be able to track and compare the results of multiple runs, and would require more skills
and steps than using Vertex AI Experiments and Vertex AI Pipelines. Vertex AI Pipelines is a
service that can orchestrate machine learning workflows using Vertex AI. Vertex AI Pipelines can
run preprocessing and training steps on custom Docker images, and evaluate, deploy, and
monitor the machine learning model.
* A custom model training job is a type of pipeline step that can train a custom model by using a
user-provided script or container. A custom model training job can accept pipeline parameters as
inputs, which can be used to control the training logic or data source. However, using the
Bayesian optimization method with F1 score as the target to maximize in the custom training step
would require writing code, implementing the optimization algorithm, and defining the objective
function. Moreover, this option would not be able to track and compare the results of multiple
runs, as Vertex AI Pipelines does not have a built-in feature for recording and displaying the
metrics, parameters, and artifacts of each run3.
* Option C: Creating a Vertex AI Workbench notebook for each of the different input datasets,
running different local training jobs with different combinations of the max tree depth and
optimizer learning rate parameters, and appending the results to a BigQuery table would not be
able to track and compare the results of multiple runs on the same platform, and would require
more skills and steps than using Vertex AI Experiments and Vertex AI Pipelines. Vertex AI
Workbench is a service that provides an integrated development environment for data science
and machine learning. Vertex AI Workbench allows users to create and run Jupyter notebooks on
Google Cloud, and access various tools and libraries for data analysis and machine learning.
However, creating a Vertex AI Workbench notebook for each of the different input datasets,
running different local training jobs with different combinations of the max tree depth and
optimizer learning rate parameters, and appending the results to a BigQuery table would require
creating multiple notebooks, writing code, setting up local environments, connecting to BigQuery,
loading and preprocessing the data, training and evaluating the model, and writing the results to a
BigQuery table. Moreover, this option would not be able to track and compare the results of
multiple runs on the same platform, as BigQuery is a separate service from Vertex AI Workbench,
and does not have a dashboard for visualizing and analyzing the metrics, parameters, and
artifacts of each run4.
References:
* Preparing for Google Cloud Certification: Machine Learning Engineer, Course 3: Production ML
Systems, Week 3: MLOps
* Google Cloud Professional Machine Learning Engineer Exam Guide, Section 1: Architecting
low-code ML solutions, 1.1 Developing ML models by using BigQuery ML
* Official Google Cloud Certified Professional Machine Learning Engineer Study Guide, Chapter
3: Data Engineering for ML, Section 3.2: BigQuery for ML
* Vertex AI Experiments
* Vertex AI Pipelines
* BigQuery ML
* Vertex AI Workbench

NEW QUESTION: 98
Your organization manages an online message board A few months ago, you discovered an
increase in toxic language and bullying on the message board. You deployed an automated text
classifier that flags certain comments as toxic or harmful. Now some users are reporting that
benign comments referencing their religion are being misclassified as abusive Upon further
inspection, you find that your classifier's false positive rate is higher for comments that reference
certain underrepresented religious groups. Your team has a limited budget and is already
overextended. What should you do?
A. Add synthetic training data where those phrases are used in non-toxic ways
B. Remove the model and replace it with human moderation.
C. Replace your model with a different text classifier.
D. Raise the threshold for comments to be considered toxic or harmful
Answer: (SHOW ANSWER)
The problem of the text classifier is that it has a high false positive rate for comments that
reference certain underrepresented religious groups. This means that the classifier is not able to
distinguish between toxic and non-toxic language when those groups are mentioned. One
possible reason for this is that the training data does not have enough examples of non-toxic
comments that reference those groups, leading to a biased model.
Therefore, a possible solution is to add synthetic training data where those phrases are used in
non-toxic ways, which can help the model learn to generalize better and reduce the false positive
rate. Synthetic data is artificially generated data that mimics the characteristics of real data, and
can be used to augment the existing data when the real data is scarce or imbalanced.
References:
* Preparing for Google Cloud Certification: Machine Learning Engineer, Course 5: Responsible
AI, Week
3: Fairness
* Google Cloud Professional Machine Learning Engineer Exam Guide, Section 4: Ensuring
solution quality, 4.4 Evaluating fairness and bias in ML models
* Official Google Cloud Certified Professional Machine Learning Engineer Study Guide, Chapter
9:
Responsible AI, Section 9.3: Fairness and Bias

NEW QUESTION: 99
You work for a gaming company that has millions of customers around the world. All games offer
a chat feature that allows players to communicate with each other in real time. Messages can be
typed in more than
20 languages and are translated in real time using the Cloud Translation API. You have been
asked to build an ML system to moderate the chat in real time while assuring that the
performance is uniform across the various languages and without changing the serving
infrastructure.
You trained your first model using an in-house word2vec model for embedding the chat
messages translated by the Cloud Translation API. However, the model has significant
differences in performance across the different languages. How should you improve it?
A. Add a regularization term such as the Min-Diff algorithm to the loss function.
B. Train a classifier using the chat messages in their original language.
C. Replace the in-house word2vec with GPT-3 or T5.
D. Remove moderation for languages for which the false positive rate is too high.
Answer: (SHOW ANSWER)
The problem with the current approach is that it relies on the Cloud Translation API to translate
the chat messages into a common language before embedding them with the in-house word2vec
model. This introduces two sources of error: the translation quality and the word2vec quality. The
translation quality may vary across different languages, depending on the availability of data and
the complexity of the grammar and vocabulary.
The word2vec quality may also vary depending on the size and diversity of the corpus used to
train it. These errors may affect the performance of the classifier that moderates the chat
messages, resulting in significant differences across the languages.
A better approach would be to train a classifier using the chat messages in their original
language, without relying on the Cloud Translation API or the in-house word2vec model. This
way, the classifier can learn the nuances and subtleties of each language, and avoid the errors
introduced by the translation and embedding processes. This would also reduce the latency and
cost of the moderation system, as it would not need to invoke the Cloud Translation API for every
message. To train a classifier using the chat messages in their original language, one could use a
multilingual pre-trained model such as mBERT or XLM-R, which can handle multiple languages
and domains. Alternatively, one could train a separate classifier for each language, using a
monolingual pre-trained model such as BERT or a custom model tailored to the specific language
and task.
References:
* Professional ML Engineer Exam Guide
* Preparing for Google Cloud Certification: Machine Learning Engineer Professional Certificate
* Google Cloud launches machine learning engineer certification
* [mBERT: Bidirectional Encoder Representations from Transformers]
* [XLM-R: Unsupervised Cross-lingual Representation Learning at Scale]
* [BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding]

NEW QUESTION: 100

A. Create a hot-encoding of words, and feed the encodings into your model.
B. Identify word embeddings from a pre-trained model, and use the embeddings in your model.
C. Sort the words by frequency of occurrence, and use the frequencies as the encodings in your
model.
D. Assign a numerical value to each word from 1 to 100,000 and feed the values as inputs in your
model.
Answer: (SHOW ANSWER)
* Option A is incorrect because creating a one-hot encoding of words, and feeding the encodings
into your model is not an efficient way to preprocess the words individually for a natural language
model. One-hot encoding is a method of representing categorical variables as binary vectors,
where each element corresponds to a category and only one element is 1 and the rest are 01.
However, this method is not suitable for high-dimensional and sparse data, such as words in a
large vocabulary, because it requires a lot of memory and computation, and does not capture the
semantic similarity or relationship between words2.
* Option B is correct because identifying word embeddings from a pre-trained model, and using
the embeddings in your model is a good way to preprocess the words individually for a natural
language model. Word embeddings are low-dimensional and dense vectors that represent the
meaning and usage of words in a continuous space3. Word embeddings can be learned from a
large corpus of text using neural networks, such as word2vec, GloVe, or BERT4. Using pre-
trained word embeddings can save time and resources, and improve the performance of the
natural language model, especially when the training data is limited or noisy5.
* Option C is incorrect because sorting the words by frequency of occurrence, and using the
frequencies as the encodings in your model is not a meaningful way to preprocess the words
individually for a natural language model. This method implies that the frequency of a word is a
good indicator of its importance or relevance, which may not be true. For example, the word "the"
is very frequent but not very informative, while the word "unicorn" is rare but more distinctive.
Moreover, this method does not capture the semantic similarity or relationship between words,
and may introduce noise or bias into the model.
* Option D is incorrect because assigning a numerical value to each word from 1 to 100,000 and
feeding the values as inputs in your model is not a valid way to preprocess the words individually
for a natural language model. This method implies an ordinal relationship between the words,
which may not be true.
For example, assigning the values 1, 2, and 3 to the words "apple", "banana", and "orange" does
not make sense, as there is no inherent order among these fruits. Moreover, this method does not
capture the
* semantic similarity or relationship between words, and may confuse the model with irrelevant or
misleading information.
References:
* One-hot encoding
* Word embeddings
* Word embedding
* Pre-trained word embeddings
* Using pre-trained word embeddings in a Keras model
* [Term frequency]
* [Term frequency-inverse document frequency]
* [Ordinal variable]
* [Encoding categorical features]

NEW QUESTION: 101


You have a functioning end-to-end ML pipeline that involves tuning the hyperparameters of your
ML model using Al Platform, and then using the best-tuned parameters for training. Hypertuning
is taking longer than expected and is delaying the downstream processes. You want to speed up
the tuning job without significantly compromising its effectiveness. Which actions should you
take?
Choose 2 answers
A. Decrease the number of parallel trials
B. Decrease the range of floating-point values
C. Set the early stopping parameter to TRUE
D. Change the search algorithm from Bayesian search to random search.
E. Decrease the maximum number of trials during subsequent training phases.
Answer: (SHOW ANSWER)
Hyperparameter tuning is the process of finding the optimal values for the parameters of a
machine learning model that affect its performance. AI Platform provides a service for
hyperparameter tuning that can run multiple trials in parallel and use different search algorithms
to find the best combination of hyperparameters.
However, hyperparameter tuning can be time-consuming and costly, especially if the search
space is large and the model training is complex. Therefore, it is important to optimize the tuning
job to reduce the time and resources required.
One way to speed up the tuning job is to set the early stopping parameter to TRUE. This means
that the tuning service will automatically stop trials that are unlikely to perform well based on the
intermediate results. This can save time and resources by avoiding unnecessary computations for
trials that are not promising. The early stopping parameter can be set in the
trainingInput.hyperparameters field of the training job request1 Another way to speed up the
tuning job is to decrease the maximum number of trials during subsequent training phases. This
means that the tuning service will use fewer trials to refine the search space after the initial
phase. This can reduce the time required for the tuning job to converge to the optimal solution.
The maximum number of trials can be set in the trainingInput.hyperparameters.maxTrials field of
the training job request1 The other options are not effective ways to speed up the tuning job.
Decreasing the number of parallel trials will reduce the concurrency of the tuning job and increase
the overall time required. Decreasing the range of floating-point values will reduce the diversity of
the search space and may miss some optimal solutions. Changing the search algorithm from
Bayesian search to random search will reduce the efficiency of the tuning job and may require
more trials to find the best solution1 References: 1: Hyperparameter tuning overview

NEW QUESTION: 102


You recently developed a wide and deep model in TensorFlow. You generated training datasets
using a SQL script that preprocessed raw data in BigQuery by performing instance-level
transformations of the data. You need to create a training pipeline to retrain the model on a
weekly basis. The trained model will be used to generate daily recommendations. You want to
minimize model development and training time. How should you develop the training pipeline?
A. Use the Kubeflow Pipelines SDK to implement the pipeline Use the BigQueryJobop
component to run the preprocessing script and the customTrainingJobop component to launch a
Vertex Al training job.
B. Use the Kubeflow Pipelines SDK to implement the pipeline. Use the
dataflowpythonjobopcomponent to preprocess the data and the customTraining JobOp
component to launch a Vertex Al training job.
C. Use the TensorFlow Extended SDK to implement the pipeline Use the Examplegen component
with the BigQuery executor to ingest the data the Transform component to preprocess the data,
and the Trainer component to launch a Vertex Al training job.
D. Use the TensorFlow Extended SDK to implement the pipeline Implement the preprocessing
steps as part of the input_fn of the model Use the ExampleGen component with the BigQuery
executor to ingest the data and the Trainer component to launch a Vertex Al training job.
Answer: (SHOW ANSWER)
* Explanation: TensorFlow Extended (TFX) is a platform for building end-to-end machine learning
pipelines using TensorFlow. TFX provides a set of components that can be orchestrated using
either the TFX SDK or Kubeflow Pipelines. TFX components can handle different aspects of the
pipeline, such as
* data ingestion, data validation, data transformation, model training, model evaluation, model
serving, and more. TFX components can also leverage other Google Cloud services, such as
BigQuery, Dataflow, and Vertex AI.
* Why not A: Using the Kubeflow Pipelines SDK to implement the pipeline is a valid option, but
using the BigQueryJobOp component to run the preprocessing script is not optimal. This would
require writing and maintaining a separate SQL script for data transformation, which could
introduce inconsistencies and errors. It would also make it harder to reuse the same
preprocessing logic for both training and serving.
* Why not B: Using the Kubeflow Pipelines SDK to implement the pipeline is a valid option, but
using the DataflowPythonJobOp component to preprocess the data is not optimal. This would
require writing and maintaining a separate Python script for data transformation, which could
introduce inconsistencies and errors. It would also make it harder to reuse the same
preprocessing logic for both training and serving.
* Why not D: Using the TensorFlow Extended SDK to implement the pipeline is a valid option, but
implementing the preprocessing steps as part of the input_fn of the model is not optimal. This
would make the preprocessing logic tightly coupled with the model code, which could reduce
modularity and flexibility. It would also make it harder to reuse the same preprocessing logic for
both training and serving.

NEW QUESTION: 103


You are building a linear regression model on BigQuery ML to predict a customer's likelihood of
purchasing your company's products. Your model uses a city name variable as a key predictive
component. In order to train and serve the model, your data must be organized in columns. You
want to prepare your data using the least amount of coding while maintaining the predictable
variables. What should you do?
A. Create a new view with BigQuery that does not include a column with city information
B. Use Dataprep to transform the state column using a one-hot encoding method, and make each
city a column with binary values.
C. Use Cloud Data Fusion to assign each city to a region labeled as 1, 2, 3, 4, or 5r and then use
that number to represent the city in the model.
D. Use TensorFlow to create a categorical variable with a vocabulary list Create the vocabulary
file, and upload it as part of your model to BigQuery ML.
Answer: (SHOW ANSWER)
One-hot encoding is a technique that converts categorical variables into numerical variables by
creating dummy variables for each possible category. Each dummy variable has a value of 1 if
the original variable belongs to that category, and 0 otherwise1. One-hot encoding can help linear
regression models to capture the effect of different categories on the target variable without
imposing any ordinal relationship among them2.
Dataprep is a service that allows you to explore, clean, and transform your data for analysis and
machine learning. You can use Dataprep to apply one-hot encoding to your city name variable
and make each city a column with binary values3. This way, you can prepare your data using the
least amount of coding while maintaining the predictive variables. Therefore, using Dataprep to
transform the state column using a one-hot encoding method is the best option for this use case.
References:
* One Hot Encoding: A Beginner's Guide
* One-Hot Encoding in Linear Regression Models
* Dataprep documentation

NEW QUESTION: 104


You work on the data science team at a manufacturing company. You are reviewing the
company's historical sales data, which has hundreds of millions of records. For your exploratory
data analysis, you need to calculate descriptive statistics such as mean, median, and mode;
conduct complex statistical tests for hypothesis testing; and plot variations of the features over
time You want to use as much of the sales data as possible in your analyses while minimizing
computational resources. What should you do?
A. Spin up a Vertex Al Workbench user-managed notebooks instance and import the dataset Use
this data to create statistical and visual analyses
B. Visualize the time plots in Google Data Studio. Import the dataset into Vertex Al Workbench
user-managed notebooks Use this data to calculate the descriptive statistics and run the
statistical analyses
C. Use BigQuery to calculate the descriptive statistics. Use Vertex Al Workbench user-managed
notebooks to visualize the time plots and run the statistical analyses.
D Use BigQuery to calculate the descriptive statistics, and use Google Data Studio to visualize
the time plots. Use Vertex Al Workbench user-managed notebooks to run the statistical analyses.
Answer: C (LEAVE A REPLY)
The best option for analyzing large and complex datasets while minimizing computational
resources is to use a combination of BigQuery and Vertex AI Workbench. BigQuery is a
serverless, scalable, and cost-effective data warehouse that can perform fast and interactive
queries on petabytes of data. BigQuery can calculate descriptive statistics such as mean, median,
and mode by using SQL functions such as AVG, PERCENTILE_CONT, and MODE. Vertex AI
Workbench is a managed service that provides an integrated development environment for data
science and machine learning. Vertex AI Workbench allows users to create and run Jupyter
notebooks on Google Cloud, and access various tools and libraries for data visualization and
statistical analysis. Vertex AI Workbench can connect to BigQuery and use the results of the
queries to create time plots and run statistical tests for hypothesis testing. By using BigQuery and
Vertex AI Workbench, users can leverage the power and flexibility of Google Cloud to perform
exploratory data analysis on large and complex datasets. References:
* Preparing for Google Cloud Certification: Machine Learning Engineer, Course 2: Data
Engineering for ML on Google Cloud, Week 1: Introduction to Data Engineering for ML
* Google Cloud Professional Machine Learning Engineer Exam Guide, Section 1: Architecting
low-code ML solutions, 1.1 Developing ML models by using BigQuery ML
* Official Google Cloud Certified Professional Machine Learning Engineer Study Guide, Chapter
3: Data Engineering for ML, Section 3.2: BigQuery for ML

NEW QUESTION: 105


You work for an online grocery store. You recently developed a custom ML model that
recommends a recipe when a user arrives at the website. You chose the machine type on the
Vertex Al endpoint to optimize costs by using the queries per second (QPS) that the model can
serve, and you deployed it on a single machine with 8 vCPUs and no accelerators.
A holiday season is approaching and you anticipate four times more traffic during this time than
the typical daily traffic You need to ensure that the model can scale efficiently to the increased
demand. What should you do?
A. 1, Maintain the same machine type on the endpoint.
2 Set up a monitoring job and an alert for CPU usage
3 If you receive an alert add a compute node to the endpoint
B. 1 Change the machine type on the endpoint to have 32 vCPUs
2. Set up a monitoring job and an alert for CPU usage
3 If you receive an alert, scale the vCPUs further as needed
C. 1 Maintain the same machine type on the endpoint Configure the endpoint to enable
autoscalling based on vCPU usage.
2 Set up a monitoring job and an alert for CPU usage
3 If you receive an alert investigate the cause
D. 1 Change the machine type on the endpoint to have a GPU_ Configure the endpoint to enable
autoscaling based on the GPU usage.
2 Set up a monitoring job and an alert for GPU usage.
3 If you receive an alert investigate the cause.
Answer: (SHOW ANSWER)
Vertex AI Endpoint is a service that allows you to serve your ML models online and scale them
automatically.
You can use Vertex AI Endpoint to deploy the custom ML model that you developed for
recommending recipes to the users. You can maintain the same machine type on the endpoint,
which is a single machine with
8 vCPUs and no accelerators. This machine type can optimize the costs by using the queries per
second (QPS) that the model can serve. You can also configure the endpoint to enable
autoscaling based on vCPU usage.
Autoscaling is a feature that allows the endpoint to adjust the number of compute nodes based on
the traffic demand. By enabling autoscaling based on vCPU usage, you can ensure that the
endpoint can scale efficiently to the increased demand during the holiday season, without
overprovisioning or underprovisioning the resources. You can also set up a monitoring job and an
alert for CPU usage. Monitoring is a service that allows you to collect and analyze the metrics and
logs from your Google Cloud resources. You can use Monitoring to monitor the CPU usage of
your endpoint, which is an indicator of the load and performance of your model.
You can also set up an alert for CPU usage, which is a feature that allows you to receive
notifications when the CPU usage exceeds a certain threshold. By setting up a monitoring job and
an alert for CPU usage, you can keep track of the health and status of your endpoint, and detect
any issues or anomalies. If you receive an alert, you can investigate the cause by using the
Monitoring dashboard, which provides a graphical interface for viewing and analyzing the metrics
and logs from your endpoint. You can also use the Monitoring dashboard to troubleshoot and
resolve the issues, such as adjusting the autoscaling parameters, optimizing the model, or
updating the machine type. By using Vertex AI Endpoint, autoscaling, and Monitoring, you can
ensure that the model can scale efficiently to the increased demand during the holiday season,
and handle any issues or alerts that might arise. References:
* [Vertex AI Endpoint documentation]
* [Autoscaling documentation]
* [Monitoring documentation]
* [Preparing for Google Cloud Certification: Machine Learning Engineer Professional Certificate]

NEW QUESTION: 106


You recently used BigQuery ML to train an AutoML regression model. You shared results with
your team and received positive feedback. You need to deploy your model for online prediction as
quickly as possible. What should you do?
A. Retrain the model by using BigQuery ML. and specify Vertex Al as the model registry Deploy
the model from Vertex Al Model Registry to a Vertex Al endpoint.
B. Retrain the model by using Vertex Al Deploy the model from Vertex Al Model Registry to a
Vertex Al endpoint.
C. Alter the model by using BigQuery ML and specify Vertex Al as the model registry Deploy the
model from Vertex Al Model Registry to a Vertex Al endpoint.
D. Export the model from BigQuery ML to Cloud Storage Import the model into Vertex Al Model
Registry Deploy the model to a Vertex Al endpoint.
Answer: (SHOW ANSWER)
BigQuery ML is a service that allows you to create and train ML models using SQL queries. You
can use BigQuery ML to train an AutoML regression model, which is a type of model that
automatically selects the best features and architecture for your data. You can also specify Vertex
AI as the model registry, which is a service that allows you to store and manage your ML models.
By using Vertex AI as the model registry, you can easily deploy your model to a Vertex AI
endpoint, which is a service that allows you to serve your ML models online and scale them
automatically. By using BigQuery ML, Vertex AI model registry, and Vertex AI endpoint, you can
deploy your model for online prediction as quickly as possible, without having to export, import, or
retrain your model. References:
* BigQuery ML documentation
* Vertex AI documentation
* Preparing for Google Cloud Certification: Machine Learning Engineer Professional Certificate

Valid Professional-Machine-Learning-Engineer Dumps shared by ExamDiscuss.com for


Helping Passing Professional-Machine-Learning-Engineer Exam! ExamDiscuss.com now offer
the newest Professional-Machine-Learning-Engineer exam dumps, the ExamDiscuss.com
Professional-Machine-Learning-Engineer exam questions have been updated and answers
have been corrected get the newest ExamDiscuss.com Professional-Machine-Learning-
Engineer dumps with Test Engine here:
https://www.examdiscuss.com/Google/exam/Professional-Machine-Learning-
Engineer/premium/ (273 Q&As Dumps, 35%OFF Special Discount Code: freecram)

NEW QUESTION: 107


You need to train a regression model based on a dataset containing 50,000 records that is stored
in BigQuery.
The data includes a total of 20 categorical and numerical features with a target variable that can
include negative values. You need to minimize effort and training time while maximizing model
performance. What approach should you take to train this regression model?
A. Create a custom TensorFlow DNN model.
B. Use BQML XGBoost regression to train the model
C. Use AutoML Tables to train the model without early stopping.
D. Use AutoML Tables to train the model with RMSLE as the optimization objective
Answer: (SHOW ANSWER)
AutoML Tables is a service that allows you to automatically build, analyze, and deploy machine
learning models on tabular data. It is suitable for large-scale regression and classification
problems, and it supports various optimization objectives, data splitting methods, and
hyperparameter tuning algorithms. AutoML Tables can handle both categorical and numerical
features, and it can also handle missing values and outliers.
AutoML Tables is a good choice for this problem because it minimizes the effort and training time
required to train a regression model, while maximizing the model performance.
RMSLE stands for Root Mean Squared Logarithmic Error, and it is a metric that measures the
average difference between the logarithm of the predicted values and the logarithm of the actual
values. RMSLE is useful for regression problems where the target variable can include negative
values, and where large differences between small values are more important than large
differences between large values. For example, RMSLE penalizes underestimating a value of 10
by 2 more than overestimating a value of 1000 by
20. RMSLE is a good optimization objective for this problem because it can handle negative
values in the target variable, and it can reduce the impact of outliers and large errors.
For more information about AutoML Tables and RMSLE, see the following references:
* AutoML Tables: end-to-end workflows on AI Platform Pipelines
* Predict workload failures before they happen with AutoML Tables
* How to Calculate RMSE in R

NEW QUESTION: 108


You are a data scientist at an industrial equipment manufacturing company. You are developing a
regression model to estimate the power consumption in the company's manufacturing plants
based on sensor data collected from all of the plants. The sensors collect tens of millions of
records every day. You need to schedule daily training runs for your model that use all the data
collected up to the current date. You want your model to scale smoothly and require minimal
development work. What should you do?
A. Develop a custom TensorFlow regression model, and optimize it using Vertex Al Training.
B. Develop a regression model using BigQuery ML.
C. Develop a custom scikit-learn regression model, and optimize it using Vertex Al Training
D. Develop a custom PyTorch regression model, and optimize it using Vertex Al Training
Answer: (SHOW ANSWER)
BigQuery ML is a powerful tool that allows you to build and deploy machine learning models
directly within BigQuery, Google's fully-managed, serverless data warehouse. It allows you to
create regression models using SQL, which is a familiar and easy-to-use language for many data
scientists. It also allows you to scale smoothly and require minimal development work since you
don't have to worry about cluster management and it's fully-managed by Google.
BigQuery ML also allows you to run your training on the same data where it's stored, this will
minimize data movement, and thus minimize cost and time.
References:
* BigQuery ML
* BigQuery ML for regression
* BigQuery ML for scalability

NEW QUESTION: 109


You built a custom ML model using scikit-learn. Training time is taking longer than expected. You
decide to migrate your model to Vertex AI Training, and you want to improve the model's training
time. What should you try out first?
A. Migrate your model to TensorFlow, and train it using Vertex AI Training.
B. Train your model in a distributed mode using multiple Compute Engine VMs.
C. Train your model with DLVM images on Vertex AI, and ensure that your code utilizes NumPy
and SciPy internal methods whenever possible.
D. Train your model using Vertex AI Training with GPUs.
Answer: (SHOW ANSWER)
* Option A is incorrect because migrating your model to TensorFlow, and training it using Vertex
AI Training, is not the easiest way to improve the model's training time. TensorFlow is a
framework that allows you to create and train ML models using Python or other languages. Vertex
AI Training is a service that allows you to train and optimize ML models using built-in algorithms
or custom containers.
However, this option requires significant code changes, as TensorFlow and scikit-learn have
different APIs and functionalities. Moreover, this option does not leverage the parallelism or the
scalability of the cloud, as it only uses a single instance.
* Option B is incorrect because training your model in a distributed mode using multiple Compute
Engine VMs, is not the most convenient way to improve the model's training time. Compute
Engine is a service that allows you to create and manage virtual machines that run on Google
Cloud. You can use Compute Engine to run your scikit-learn model in a distributed mode, by
using libraries such as Dask or Joblib.
However, this option requires more effort and resources than option D, as it involves creating and
configuring the VMs, installing and maintaining the libraries, and writing and running the
distributed code.
* Option C is incorrect because training your model with DLVM images on Vertex AI, and
ensuring that your code utilizes NumPy and SciPy internal methods whenever possible, is not the
most effective way to improve the model's training time. DLVM (Deep Learning Virtual Machine)
images are preconfigured VM images that include popular ML frameworks and tools, such as
TensorFlow, PyTorch, or scikit-learn1. You can use DLVM images on Vertex AI to train your
scikit-learn model, by using a custom container. NumPy and SciPy are libraries that provide
numerical and scientific computing functionalities for Python. You can use NumPy and SciPy
internal methods to optimize your scikit-learn code, as they are faster and more efficient than pure
Python code2. However, this option does not leverage the parallelism or the scalability of the
cloud, as it only uses a single instance. Moreover, this option may not have a significant impact
on the training time, as scikit-learn already relies on NumPy and SciPy for most of its operations3.
* Option D is correct because training your model using Vertex AI Training with GPUs, is the best
way to improve the model's training time. A GPU (Graphics Processing Unit) is a hardware
accelerator that can perform parallel computations faster than a CPU (Central Processing Unit)4.
Vertex AI Training is a service that allows you to train and optimize ML models using built-in
algorithms or custom containers. You can use Vertex AI Training with GPUs to train your scikit-
learn model, by using a custom container and specifying the accelerator type and count5. By
using Vertex AI Training with GPUs, you can leverage the parallelism and the scalability of the
cloud, and speed up the training process significantly, without changing your code.
References:
* DLVM images
* NumPy and SciPy
* scikit-learn dependencies
* GPU overview
* Vertex AI Training with GPUs
* [scikit-learn overview]
* [TensorFlow overview]
* [Compute Engine overview]
* [Dask overview]
* [Joblib overview]
* [Vertex AI Training overview]

Valid Professional-Machine-Learning-Engineer Dumps shared by ExamDiscuss.com for


Helping Passing Professional-Machine-Learning-Engineer Exam! ExamDiscuss.com now offer
the newest Professional-Machine-Learning-Engineer exam dumps, the ExamDiscuss.com
Professional-Machine-Learning-Engineer exam questions have been updated and answers
have been corrected get the newest ExamDiscuss.com Professional-Machine-Learning-
Engineer dumps with Test Engine here:
https://www.examdiscuss.com/Google/exam/Professional-Machine-Learning-
Engineer/premium/ (273 Q&As Dumps, 35%OFF Special Discount Code: freecram)

You might also like