0% found this document useful (0 votes)
249 views19 pages

GCP 5

The document provides answers to multiple choice questions about various Google Cloud Platform services and best practices. The questions cover topics such as making on-premises data accessible to Cloud Dataflow, improving application availability and reducing costs, managing access to resources across projects, optimizing storage costs, and designing secure access between services in different projects.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
249 views19 pages

GCP 5

The document provides answers to multiple choice questions about various Google Cloud Platform services and best practices. The questions cover topics such as making on-premises data accessible to Cloud Dataflow, improving application availability and reducing costs, managing access to resources across projects, optimizing storage costs, and designing secure access between services in different projects.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 19

1.

Your company has a massive quantity of unstructured data in text, Apache


AVRO and PARQUET files in the on-premise data centre and wants to
transform this data using a Dataflow job and migrate cleansed/enriched data
to BigQuery. How should you make the on-premise files accessible to Cloud
Dataflow?

A. Migrate the data from the on-premises data centre to Cloud Spanner by
using the upload files function.
B. Migrate the data from the on-premises data centre to Cloud SQL for
MySQL by using the upload files function.
C. Migrate the data from the on-premises data centre to Cloud Storage by
using a custom script with gsutil commands.
D. Migrate the data from the on-premises data centre to BigQuery by using a
custom script with bq commands.

Answer: C

2. Your business-critical application deployed on a compute engine instance


in us-west 1-a zone suffered an outage due to GCP zone failure. You want to
modify the application to be immune to zone failures while minimizing costs.
What should you do?

A. Direct the traffic through a Global HTTP(s) Load Balancer to shield your
application from GCP zone failures.
B. Provision another compute engine instance in us-west1-b and balance the
traffic across both zones.
C. Ensure you have hourly snapshots of the disk in Google Cloud Storage. In
the unlikely event of a zonal outage, use the snapshots to provision a new
Compute Engine Instance in a different zone.
D. Replace the single instance with a Managed Instance Group (MIG) and
autoscaling enabled. Configure a health check to detect failures rapidly.

Answer: B

3. You deployed a mission-critical application on Google Compute Engine.


Your operations team have asked you to enable measures to prevent
engineers from accidentally destroying the instance. What should you do?

A. Turn on deletion protection on the compute engine instance.


B. Uncheck “Delete boot disk when instance is deleted” option when
provisioning the compute engine instance.
C. Deploy the application on a preemptible compute engine instance.
D. Enable automatic restart on the instance.

Answer: A

4. Your production applications are distributed across several Google Cloud


Platform (GCP) projects, and your operations team want to efficiently manage
all the production projects and applications using gcloud SDK on Cloud Shell.
What should you recommend they do to achieve this in the fewest possible
steps?

A. Create a gcloud configuration for each production project. To manage


resources of a particular project, run gcloud init to update and initialize the
relevant gcloud configuration.
B. Use the default gcloud configuration on cloud shell. To manage resources
of a particular project, activate the relevant gcloud configuration.
C. Create a gcloud configuration for each production project. To manage
resources of a particular project, activate the relevant gcloud configuration.
D. Use the default gcloud configuration on cloud shell. To manage resources
of a particular project, run gcloud init to update and initialize the relevant
gcloud configuration.

Answer: C

5. You want to optimize the storage costs for long term archival of logs. Logs
are accessed frequently in the first 30 days and only retrieved after that if
there is any special requirement in the annual audit. The auditors may need to
look into log entries of the previous three years. What should you do?

A. Store the logs in Nearline Storage Class and set up a lifecycle policy to
transition the files older than 30 days to Archive Storage Class.
B. Store the logs in Standard Storage Class and set up a lifecycle policy to
transition the files older than 30 days to Archive Storage Class.
C. Store the logs in Standard Storage Class and set up lifecycle policies to
transition the files older than 30 days to Coldline Storage Class, and files
older than 1 year to Archive Storage Class.
D. Store the logs in Nearline Storage Class and set up lifecycle policies to
transition the files older than 30 days to Coldline Storage Class, and files
older than 1 year to Archive Storage Class.

Answer: B

6. Your company runs most of its compute workloads in Google Compute


Engine in the europe-west1-b zone. Your operations team use Cloud Shell to
manage these instances. They want to know if it is possible to designate a
default compute zone and not supply the zone parameter when running each
command in the CLI. What should you do?

A. In GCP Console set europe-west1-! zone in the default location in Compute


Engine Settings.
B. Run gcloud config to set europe-west1-b as the default zone.
C. Update the gcloud configuration file ~/config.ini to set europe-west1-b as
the default zone.
D. Add a metadata entry in the Compute Engine Settings page with key:
compute/zone and value: europe-west1-b.

Answer: A
7. You run a business-critical application in a Google Cloud Compute Engine
instance, and you want to set up a cost-efficient solution for backing up the
data on the boot disk. You want a solution that:
minimizes operational overhead
– backs up boot disks daily
. allows quick restore of the backups when needed, e.g. disaster scenarios
. deletes backups older than a month automatically.
What should you do?

A. Enable Snapshot Schedule on the disk to enable automated snapshots


per schedule.
B. Deploy a Cloud Function to initiate the creation of instance templates for
all instances daily.
C. Configure a Cloud Task to initiate the creation of images for all instances
daily and upload them to Cloud Storage.
D. Set up a cron job with a custom script that uses gcloud APIs to create a
new disk from existing instance disk for all instances daily.

Answer: A

8. The operations manager has asked you to identify the IAM users with
Project Editor role on the GCP production project. What should you do?

A. Turn on IAM Audit logging and build a Cloud Monitoring dashboard to


display this information.
B. Extract all project-wide SSH keys.
C. Execute gcloud projects get-iam-policy to retrieve this information.
D. Check the permissions assigned in all Identity Aware Proxy (IAP) tunnels.

Answer: C

9. Your company’s compute workloads are split between the on-premises


data centre and Google Cloud Platform. The on-premises data centre is
connected to Google Cloud network by Cloud VPN. You have a requirement
to provision a new non-publicly-reachable compute engine instance on a c2-
standard-8 machine type in australia-southeast1- zone. What should you do?

A. Configure a route to route all traffic to the public IP of compute engine


instance through the VPN tunnel.
B. Provision the instance without a public IP address.
C. Provision the instance in a subnet that has Google Private Access enabled.
D. Provision the instance in a subnetwork that has all egress traffic disabled.

Answer: B

10. You manage an overnight batch job that uses 20 VMs to transfer
customer information from a CRM system to BigQuery dataset. The job can
tolerate some VMs going down. The current high cost of the VMs make the
overnight job not viable, and you want to reduce the costs. What should you
do?
A. Use preemptible compute engine instances to reduce cost.
B. Use a fleet of f1-micro instances behind a Managed Instances Group (MIG)
with autoscaling. Set minimum and maximum nodes to 20.
C. Use tiny f1-micro instances to reduce cost.
D. Use a fleet of f1-micro instances behind a Managed Instances Group (MIG)
with autoscaling and minimum nodes set to 1.

Answer: A

11. You plan to deploy an application to Google Compute Engine instance,


and it relies on making connections to a Cloud SQL database for retrieving
information about book publications. To minimize costs, you are developing
this application on your local workstation, and you want it to connect to a
Cloud SQL instance. Your colleague suggested setting up Application Default
Credentials on your workstation to make the transition to Google Cloud easier.
You are now ready to move the application to Google Compute Engine
instance. You want to follow Google recommended practices to enable secure
IAM access. What should you do?

A. Grant the necessary IAM roles to a service account, download the JSON
key file and package it with your application.
B. Grant the necessary IAM roles to the service account used by Google
Compute Engine instance.
C. Grant the necessary IAM roles to a service account and configure the
application running on Google Compute Engine instance to use this service
account.
D. Grant the necessary IAM roles to a service account, store its credentials in
a config file and package it with your application.

Answer: B

12. You want to run an application in Google Compute Engine in the app-tier
GCP project and have it export data from Cloud Bigtable to daily-us-customer-
export Cloud Storage bucket in the data-warehousing project. You plan to run
a Cloud Dataflow job in the data-arehousing project to pick up data from this
bucket for further processing. How should you design the IAM access to
enable the compute engine instance push objects to daily-us-customer-export
Cloud Storage bucket in the data-warehousing project?

A. Update the access control on daily-us-customer-export Cloud Storage


bucket to make it public. Create a subfolder inside the bucket with a
randomized name and have the compute engine instance push objects to this
folder.
B. Ensure both the projects are in the same GCP folder in the resource
hierarchy.
C. Grant the service account used by the compute engine in app-tier GCP
project roles/storage.objectCreator IAM role on the daily-us-customer-export
Cloud Storage bucket.
D. Grant the service account used by the compute engine in app-tier GCP
project roles/storage.objectCreator IAM role on app-tier GCP project.

Answer: C

13. You run a batch job every month in your on-premises data centre that
downloads clickstream logs from Google Cloud Storage bucket, enriches the
data and stores them in Cloud BigTable. The job runs for 32 hours on average,
can be restarted if interrupted, and must complete. You want to migrate this
batch job onto a cost-efficient GCP compute service. How should you deploy
it?

A. Deploy the batch job in a GKE Cluster with preemptible VM node pool.
B. Deploy the batch job on a fleet of Google Cloud Compute Engine
preemptible VM in a Managed Instances Group (MIG) with autoscaling.
C. Deploy the batch job on a Google Cloud Compute Engine Preemptible
VM.
D. Deploy the batch job on a Google Cloud Compute Engine non-
preemptible VM. Restart instances as required.

Answer: A

14. Your gaming backend uses Cloud Spanner to store leaderboard and
player profile data. You want to scale the spanner instances based on
predictable usage patterns. What should you do?

A. Configure alerts in Cloud Monitoring to alert Google Operations Support


team and have them use their scripts to scale up or scale down the spanner
instance as necessary.
B. Configure a Cloud Scheduler job to invoke a Cloud Function that reviews
the relevant Cloud Monitoring metrics and resizes the Spanner instance as
necessary.
C. Configure alerts in Cloud Monitoring to trigger a Cloud Function via
webhook, and have the Cloud Function scale up or scale down the spanner
instance as necessary.
D. Configure alerts in Cloud Monitoring to alert your operations team and
have them manually scale up or scale down the spanner instance as
necessary.

Answer: C

15. You are running an application on a Google Compute Engine instance.


You want tocreate multiple copies of this VM to handle the burst in traffic.
What should you do?

A. Create a snapshot of the compute engine instance disk and create


images from this snapshot to handle the burst in traffic.
B. Create a snapshot of the compute engine instance disk and create
instances from this snapshot to handle the burst in traffic.
C. Create a snapshot of the compute engine instance disk, create a custom
image from the snapshot, create instances from this image to handle the burst
in traffic.
D. Create a snapshot of the compute engine instance disk, create custom
images from the snapshot to handle the burst in traffic.

Answer: C

16. You migrated a mission-critical application from the on-premises data


centre to Google Kubernetes Engine (GKE) which uses e2-standard-2
machine types. You want to deploy additional pods on c2-standard-16
machine types. How can you do this without causing application downtime?

A. Run gcloud container clusters upgrade to move to c2-standard– 16


machine types. Terminate all existing pods.
B. Create a new GKE cluster with node pool instances of type c2-
standard- – 16. Deploy the application on the new GKE cluster and delete the
old GKE Cluster.
C. Create a new GKE cluster with two node pools – one with e2-standard-2
machine types and other with c2-standard-16 machine types. Deploy the
application on the new GKE cluster and delete the old GKE Cluster.
D. Update the existing cluster to add a new node pool with c2-standard-16
machine types and deploy the pods.

Answer: C

17. You are migrating a complex on-premises data warehousing solution to


Google Cloud. You plan to create a fleet of Google Compute Engine
instances behind a Managed Instances Group (MIG) in the app-tier project,
and BigQuery in the data-warehousing project. How should you configure the
service accounts used by Compute Engine instances to allow them query
access to BigQuery datasets?

A. Grant the compute engine service account roles/bigquery.dataViewer


role on the data-warehousing GCP project.
B. Grant the compute engine service account roles/owner on data-
warehousing GCP project.
C. Grant the compute engine service account roles/owner on data-
warehousing GCP project and roles/biqquery.dataViewer role on the app-tier
GCP project.
D. Grant the BigQuery service account roles/owner on app-tier GCP project.

Answer: A

18. You deployed the Finance teams’ Payroll application to Google Compute
Engine, and this application is used by staff during regular business hours.
The operations team want to backup the VMs daily outside the business
hours and delete images older than 50 days to save costs. They need an
automated solution with the least operational overhead and the least number
of GCP services. What should they do?

A. Navigate to the Compute Engine Disk section of your VM instance in the


GCP console and enable a snapshot schedule for automated creation of daily
snapshots. Set Auto-Delete snapshots after to 50 days.
B. Add a metadata tag on the Google Compute Engine instance to enable
snapshot creation. Add a second metadata tag to specify the snapshot
schedule, and a third metadata tag to specify the retention period.
C. Use Cloud Scheduler to trigger a Cloud Function that creates snapshots
of the disk daily. Use Cloud Scheduler to trigger another Cloud Function that
iterates over the snapshots and deletes snapshots older than 50 days.
D. Use AppEngine Cron service to trigger a custom script that creates
snapshots of the disk daily. Use AppEngine Cron service to trigger another
custom script that iterates over the snapshots and deletes snapshots older
than 50 days.

Answer: A

19. Your compliance team wants to review the audit logs and data access
logs in the production GCP project. You want to follow Google recommended
practices. What should you do?

A. Grant the compliance team a custom IAM role that has


logging.privateLogEntries. list permission. Let the compliance team know they
can also query IAM policy changes in Cloud Logging.
B. Export logs to Cloud Storage and grant the compliance team a custom IAM
role that has logging.privateLogEntries.list permission.
C. Export logs to Cloud Storage and grant the compliance team
roles/logging.privateLogViewer lAMrole.
D. Grant the compliance team roles/logging.privateLogViewer\AM role. Let the
compliance team know they can also query IAM policy changes in Cloud
Logging.

Answer: D

20. You want to migrate a legacy application from your on-premises data
centre to Google Cloud Platform. The application serves SSL encrypted traffic
from worldwide clients on TCP port 443. What GCP Loadbalancing service
should you use to minimize latency for all clients?
A. External HTTP(S) Load Balancer.

B. Internal TCP/UDP Load Balancer.

C. Network TCP/UDP Load Balancer.

D. SSL Proxy Load Balancer.

Answer: D

21. You are deploying an application on the Google Compute Engine, and you
want to minimize network egress costs. The organization has a policy that
requires you to block all but essential egress traffic. What should you do?

A. Enable a firewall rule at priority 100 to allow essential egress traffic.

B. Enable a firewall rule at priority 100 to allow ingress and essential egress
traffic.

C. Enable a firewall rule at priority 100 to block all egress traffic, and another
firewall rule at priority 65534 to allow essential egress traffic.

D. Enable a firewall rule at priority 65534 to block all egress traffic, and
another firewall rule at priority 100 to allow essential egress traffic.

Answer: D

22. You work for a startup company where every developer has a dedicated
development GCP project linked to a central billing account. Your finance lead
is concerned that some developers may leave some services running
unnecessarily or may not understand the cost implications of turning on
specific services in Google Cloud Platform. They want to be alerted when a
developer spends more than 750$ per month in their GCP project. What
should you do?
A. Export Billing data from each development GCP projects to a separate
BigQuery dataset. On each dataset, use a Data Studio dashboard to plot the
spending.

B. Set up a budget for each development GCP projects. For each budget,
trigger an email notification when the spending exceeds $750.

C. Export Billing data from all development GCP projects to a single BigQuery
dataset. Use a Data Studio dashboard to plot the spend.

D. Set up a single budget for all development GCP projects. Trigger an email
notification when the spending exceeds $750 in the budget.

Answer: B

23. Your company has deployed all its production applications in a single
Google Cloud Project and uses several GCP projects for development and
test environments. The operations team requires access to all production
services in this project to debug live issues and deploy enhancements. Your
security team prevents the creation of IAM roles that automatically broaden to
include new permissions/services in future. How should you design the IAM
role for operations team?

A. Create a custom role with the necessary permissions and grant the role on
the production GCP project to all members of the operations team.

B. Grant the Project Editor role at the organization level to all members of the
operations team.

C. Grant the Project Editor role on the production GCP project to all members
of the operations team.

D. Create a custom role with the necessary permissions and grant the role at
the organization level to all members of the operations team.

Answer: A
24. EU GDPR requires you to respond to a Subject Access Request (SAR)
within one month. To be compliant, your company deployed an application
that uses Apache WebServer to provide SAR archive (tar) files back to
customers requesting them. Your compliance team has asked you to send
them an email notification when the network egress charges for this server in
the GCP project exceeds 250 dollars per month. What should you do?

A. Export the logs from Apache server to Cloud Logging and deploy a Cloud
Function to parse the logs, extract and sum up the size of response payload
for all requests during the current month; and send an email notification when
spending exceeds $250.

B. Configure a budget with the scope set to the billing account, the amount set
to $250, threshold rule set to 100% of actual cost & trigger email notifications
when spending exceeds the threshold.

C. Export the project billing data to a BigQuery dataset and deploy a Cloud
Function to extract and sum up the network egress costs from the BigQuery
dataset for the Apache server for the current month, and send an email
notification when spending exceeds $250.

D. Configure a budget with the scope set to the project, the amount set to
$250, threshold rule set to 100% of actual cost & trigger email notifications
when spending exceeds the threshold.

Answer: C

25. You are running a business-critical application in a GKE cluster in a


subnet with cluster autoscaling enabled. A massive surge in demand for your
company’s products has seen the GKE cluster node pool scale-up until there
were no more free IP addresses available for new VMs in the subnet. What
should you do to fix this issue?

A. Add a new subnet to the same region.

B. Add a secondary (alias) IP range to the existing subnet.


C. Add a new VPC and set up VPC sharing between the new and existing
VPC.

D. Expand the range of the existing subnet.

Answer: D

26. You developed an application on App Engine Service to read data from a
BigQuery dataset and convert the data to PARQUET format. The application
is using the default app-engine service account in the app-tier GCP project.
The data team owns the BigQuery dataset in the data-warehousing project.
What IAM Access should you grant to the default app-engine service account
in app-tier GCP project?

A. Grant the default app-engine service account in the app-tier GCP project
roles/bigquery. data Viewer role on the data-warehousing project.

B. Grant the service account in the data-warehousing GCP project


roles/bigquery.job User role on the app-tier project.

C. Grant the default app-engine service account in the app-tier GCP project
roles/bigquery.dataViewer role on the same project.

D. Grant the default app-engine service account in the app-tier GCP project
roles/bigquery.jobUser role on data-warehousing project.

Answer: A

27. Your company updated its business operating model recently and no
longer need the applications deployed in the data-analytics-v1 GCP project.
You want to turn off all GCP services and APIs in this project. You want to do
this efficiently using the least number of steps while following Google
recommended practices. What should you do?
A. Ask an engineer with Project Owner IAM role to identify all resources in the
project and delete them.

B. Ask an engineer with Project Owner IAM role to locate the project and shut
down.

C. Ask an engineer with Organization Administrator IAM role to identify all


resources in the project and delete them.

D. Ask an engineer with Organization Administrator IAM role to locate the


project and shut down.

Answer: B

28. You deployed an application on a general-purpose Google Cloud


Compute Engine instance that uses a persistent zonal SSD of 300 GB. The
application downloads large Apache AVRO files from Cloud Storage, retrieve
customer details from them and saves a text file on local disk for each
customer before pushing all the text files to a Google Storage Bucket. These
operations require high disk I/O, but you find that the read and write
operations on the disk are always throttled. What should you do to improve
the throughput while keeping costs to a minimum?

A. Bump up the size of its SSD persistent disk to 1 TB.

B. Replace Zonal Persistent SSD with a Local SSD.

C. Replace Zonal Persistent SSD with a Regional Persistent SSD.

D. Bump up the CPU allocated to the general-purpose Compute Engine


instance.

Answer: B

29. You want to monitor resource utilization (RAM, Disk, Network, CPU, etc.)
for all applications in development, test and production GCP projects in a
single dashboard. What should you do?

A. In Cloud Monitoring, share charts from development, test and production


GCP projects.
B. Create a Cloud Monitoring workspace in the production project and add
development and test projects to it.
C. Make use of the default Cloud Monitoring dashboards in all the projects.
D. Grant roles/monitoring.admin to development, test and production GCP
projects.

Answer: B

30. You have an application in your on-premises data centre with an API that
is triggered when a new file is created or updated in a NAS share. You want
to migrate this solution to Google Cloud Platform and have identified Cloud
Storage as the replacement service for NAS. How should you deploy the API?

A. Deploy the API on GKE cluster and use Cloud Scheduler to trigger the
API to look for files in Cloud Storage there were created or update since the
last run.
B. Trigger a Cloud Function whenever files in Cloud Storage are created or
updated.
C. Trigger a Cloud Dataflow job whenever files in Cloud Storage are created
or updated.
D. Configure Cloud Pub/Sub to capture details of files created/modified in
Cloud Storage. Deploy the API in App Engine Standard and use Cloud
Scheduler to trigger the API to fetch information from Cloud Pub/Sub.

Answer: B

31. You are running a business-critical application in a fleet of compute


engine instances behind an autoscaled Managed Instances Group (MIG). The
MIG initiated a scale-up event to keep up with the increasing incoming traffic,
but the compute engine instance failed to create. How should you debug this
issue?

A. 1. Ensure you don’t have any persistent disks with the same name as the
VM instance. 2. Ensure the disk autodelete property is turned on
(disks.autoDelete set to true). 3. Ensure instance template syntax is valid.
B. 1. Ensure instance template syntax is valid. 2. Ensure the instance
template, instance and the persistent disk names do not conflict.
C. 1. Ensure the instance template, instance and the persistent disk names
do not conflict. 2. Ensure the disk autodelete property is turned on
(disks.autoDelete set to true).
D. 1. Ensure you don’t have any persistent disks with the same name as the
VM instance. 2. Ensure instance template syntax is valid.

Answer: A

32. You work for a multinational car insurance company that specializes in
rewarding safer drivers with cheaper premiums. Your company does this by
installing black box loT devices in its 2 million insured drivers’ cars. These
devices capture driving behaviours such as acceleration/deceleration, speed
compared to speed limits, and types of driving, such as commuting on
freeway compared to commuting on surface streets etc. You expect to receive
hundreds of events per minute from every device. You need to store this data
and retrieve data consistently based on the event time, and both operations
should be atomic. How should you store this data?

A. Store the data in Cloud Storage. Have a file per loT device and append
new data to the file.
B. Store the data in Cloud Datastore. Have an entity group per device.
C. Store the data in Cloud BigTable. Have a row key based on the ingestion
timestamp.
D. Store the data in Cloud Filestore. Have a file per loT device and append
new data to the file.

Answer: C

33. Your team creates/updates the infrastructure for all production


requirements. You need to implement a new change to the current
infrastructure and want to preview the update to the rest of your team before
committing the changes. You want to follow Google- recommended practices.
What should you?

A. Clone the production environment to create a staging environment and


deploy the proposed changes to the staging environment. Execute gcloud
compute instances list to view the changes and store the results in a Google
Cloud Storage bucket.
B. Preview the updates using Deployment Manager and store the results in a
Google Cloud Storage bucket.
C. Clone the production environment to create a staging environment and
deploy the proposed changes to the staging environment. Execute gcloud
compute instances list to view the changes and store the results in a Google
Cloud Source Repository.
D. Preview the updates using Deployment Manager and store the results in a
Google Cloud Source Repository.

Answer: B

34. Your company is building a mobile application that enables users to


upload and share images with their friends. Your company places a high
value on security. Prefers minimal maintenance (no-op), and wants to
optimize costs where possible. You are designing the backend for the app
based on these requirements:
– Enable users to upload images for only 30 minutes,
– Enable users to retrieve their images and share their images with their
friends,
– Delete images older than 50 days.
You have very little time to design the solution and take it to production. What
should you do? (Choose two)

A. Have the mobile application use signed URLs to enabled time- limited
upload to Cloud Storage.
B. Use Cloud Scheduler to trigger a Cloud Function to check for objects older
than 50 days and delete them.
C. Enable lifecycle policy on the bucket to delete objects older than 50 days.
D. Write a cron script that checks for objects older than 50 days and deletes
them.
E. Have the mobile application send the images to an SFTP server.

Answer: A & C

35. An auditor requires specific access on certain GCP services in your


Cloud project. You have started working on the first version of a custom IAM
role to enable this access. You are currently testing this role in a test GCP
project. The compliance team requires this role to be production-ready, and
want you to share with them the lifecycle stage. What should you do?

A. 1. Set the custom IAM role lifecycle stage to ALPHA while you test the role
in the test GCP project. 2. Restrict the custom IAM role to use permissions
with TESTING support level.
B. 1. Set the custom IAM role lifecycle stage to BETA while you test the role in
the test GCP project. 2. Restrict the custom IAM role to use permissions with
SUPPORTED support level.
C. 1. Set the custom IAM role lifecycle stage to BETA while you test the role
in the test GCP project. 2. Restrict the custom IAM role to use permissions
with TESTING support level.
D. 1. Set the custom IAM role lifecycle stage to ALPHA while you test the role
in the test GCP project. 2. Restrict the custom IAM role to use permissions
with SUPPORTED support level.

Answer: D

36. Your company stores an export of its Customer PII data in a multi-regional
Google Cloud storage bucket. Your legal and compliance department has
asked you to record all operations/requests on the data in this bucket. What
should you do?

A. Enable the default Cloud Storage Service account exclusive access to read
all operations and record them.
B. Use the Data Loss Prevention API to record this information.
C. Use the Identity Aware Proxy API to record this information.
D. Turn on data access audit logging in Cloud Storage to record this
information.

Answer: D

37. A finance analyst at your company is suspended pending an investigation


into alleged financial misconduct. However, their Gsuite account was not
disabled immediately. Your compliance team has asked you to find out if the
suspended employee has accessed any audit logs or BigQuery datasets after
their suspension. What should you do?
A. Search for users’ Cloud Identity username (email address) as the principal
in system event logs in Cloud Logging.
B. Search for users’ Cloud Identity username (email address) as the principal
in data access logs in Cloud Logging.
C. Search for users’ service account as the principal in admin activity logs in
Cloud Logging.
D. Search for users’ service account as the principal in data access logs in
Cloud Logging.

Answer: B

38. Your company wants to migrate all compute workloads from the on-
premises data centre to Google Cloud Compute Engine. A third-party team
provides operational support for your production applications outside business
hours. Everyone at your company has a Gsuite account, but the support team
do not. How should you grant them access to the VMs?

A. Use Cloud Identity Aware Proxy (IAP) to enable SSH tunnels to the VMs
and add the third-party team as a tunnel user.
B. Set up a firewall rule to open SSH port (TCP:22) to the IP range of the
third-party team.
C. Set up a Cloud VPN tunnel between the third-party network and your
production GCP project.
D. Add all the third party teams SSH keys to the production compute engine
instances.

Answer: A

39. All departments at your company have their own Google Cloud Projects.
You got transferred into a new department that doesn’t have a project yet, and
you are ready to deploy a new application onto a Compute Engine Instance.
What should you do?

A. In the GCP Console, enable the Compute Engine API. When creating a
new instance in the console, select the checkbox to create the instance in a
new GCP project and provide the project name and ID.
B. Use gcloud commands first to create a new project, then to enable the
Compute Engine API and finally, to launch a new compute engine instance in
the project.
C. Run gcloud compute instances create with –project flag to automatically
create the new project and a compute engine instance. When prompted to
enable the Compute Engine API, select Yes.
D. In the GCP Console, enable the Compute Engine API. Run gcloud
compute instances create with – –project flag to automatically create the new
project and a compute engine instance.

Answer: B
40. You deployed an application using Apache Tomcat server on a single
Google Cloud VM. Users are complaining of intermittent issues accessing a
specific page in the application, and you want to look at the logs on the local
disk. What should you do?

A. Configure a health check on the instance to identify the issue and email
you the logs when the application experiences the issue.
B. Check logs in Cloud Logging.
C. Check logs in the Serial Console.
D. Install the Cloud Logging Agent on the VM and configure it to send logs to
Cloud Logging. Check logs in Cloud Logging.

Answer: D

41. Your company plans to migrate all applications from the on-premise data
centre to Google Cloud Platform and requires a monthly estimate of the cost
of running these applications in GCP. How can you provide this estimate?

A. For all GCP services/APIs you are planning to use, use the GCP pricing
calculator to estimate the monthly costs.
B. For all GCP services/APIs you are planning to use, capture the pricing from
the products pricing page and use an excel sheet to estimate the monthly
costs.
C. Migrate all applications to GCP and run them for a week. Use the costs
from the Billing Report page for this week to extrapolate the monthly cost of
running all applications in GCP.
D. Migrate all applications to GCP and run them for a week. Use Cloud
Monitoring to identify the costs for this week and use it to derive the monthly
cost of running all applications in GCP.

Answer: A

42. Your company uses Google Cloud for all its compute workloads. One of
the applications that you developed has passed unit testing, and you want to
use Jenkins to deploy the application in User Acceptance Testing (UAT)
environment. Your manager has asked you to automate Jenkins installation
as quickly and efficiently as possible. What should you do?

A. Deploy Jenkins on a fleet of Google Cloud Compute Engine VMs in a


Managed Instances Group (MIG) with autoscaling.
B. Deploy Jenkins on a Google Compute Engine VM.
C. Use GCP Marketplace to provision Jenkins.
D. Deploy Jenkins on a GKE Cluster.

Answer: C
43. Your company deployed its applications across hundreds of GCP projects
that use different billing accounts. The finance team is struggling to add up all
production Cloud Opex costs and has requested your assistance for
enabling/providing a single pane of glass for all costs incurred by all
applications in Google Cloud. You want to include new costs as soon as they
become available. What should you do?

A. Use Google pricing calculator for all the services used in all GCP projects
and pass the estimated cost to finance team every month.
B. Enable Billing Export from all GCP projects to BigQuery and ask the
finance team to use Google Data Studio to visualize the data.
C. Ask the finance team to check reports view section in Cloud Billing Console.
D. Use Cloud Scheduler to trigger a Cloud Function every hour. Have the
Cloud Function download the CSV from the Cost Table page and upload the
data to BigQuery. Ask the finance team to use Google Data Studio to
visualize the data.

Answer: B

44. Your production Compute workloads are running in a subnet with a range
192.168.20.128/25. A recent surge in traffic has seen the production VMs
struggle, and you want to add more VMs, but there are no free IP addresses
in the VPC. All new and old VMs need to communicate with each other. How
can you do this with the fewest steps?

A. Create a new VPC and a new subnet with IP range 192.168.21.0/24.


Enable VPC Peering between the old VPC and new VPC. Configure a custom
Route exchange.
B. Create a new VPC network and a new subnet with IP range
192.168.21.0/24. Enable VPC Peering between the old VPC and new VPC.
C. Create a new non-overlapping Alias range in the existing VPC and
Configure the VMs to use the alias range.
D. Update the subnet range to 192.168.20.0/24.

Answer: B

45. Your compliance team has asked you to set up an external auditor access
to logs from all GCP projects for the last 60 days. The auditor wants to
combine, explore and analyze the contents of the logs from all projects quickly
and efficiently. You want to follow Google Recommended practices. What
should you do?

A. Set up a Cloud Storage sink destination to export logs from all the projects
to a bucket. Configure a lifecycle rule to delete objects older than 60 days.
Ask the auditor to query logs from the bucket.
B. Set up a Cloud Scheduler job to trigger a Cloud Function that reads and
export logs from all the projects to a BigQuery dataset. Configure the table
expiration on the dataset to 60 days. Ask the auditor to query logs from the
dataset.
C. Set up a BigQuery sink destination to export logs from all the projects to a
dataset. Configure the table expiration on the dataset to 60 days. Ask the
auditor to query logs from the dataset.
D. Ask the auditor to query logs from Cloud Logging.

Answer: C

46. You work for a multinational delivery services company that uses Apache
Cassandra DB as the backend store for its delivery track and trace system.
The existing on-premises data centre is out of space. To cope with an
anticipated increase in requests in the run-up to Christmas, you want to move
this application rapidly to Google Cloud with minimal effort whilst ensuring you
can spin up multiple stacks (development, test, production) and isolate them
from each other. How can you do this?

A. Download the installation guide for Cassandra on GCP and follow the
instructions to install the database.
B. Launch Cassandra DB from Cloud Marketplace.
C. Install an instance of Cassandra DB on Google Cloud Compute Engine,
take a snapshot of this instance and use the snapshot to spin up additional
instances of Cassandra DB.
D. Install an instance of Cassandra DB on Google Cloud Compute Engine,
take a snapshot of this instance and upload to Google Cloud Storage bucket.
Every time you need a new instance of Cassandra DB, spin up a new
compute engine instance from the snapshot.

Answer: B

You might also like