0% found this document useful (0 votes)
553 views8 pages

Real Google Cloud Associate Data Practitioner Study Questions by Brady

This document contains a series of study questions for the Google Cloud Associate Data Practitioner exam, shared by Brady. Each question is multiple-choice and includes explanations for the correct answers and why other options are incorrect. The questions cover various topics related to data transfer, migration, transformation, and management within Google Cloud services.

Uploaded by

auboncoin 95
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
553 views8 pages

Real Google Cloud Associate Data Practitioner Study Questions by Brady

This document contains a series of study questions for the Google Cloud Associate Data Practitioner exam, shared by Brady. Each question is multiple-choice and includes explanations for the correct answers and why other options are incorrect. The questions cover various topics related to data transfer, migration, transformation, and management within Google Cloud services.

Uploaded by

auboncoin 95
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 8

Real Google Cloud Associate Data Practitioner Study Questions By Brady - Page 1

Free Questions for Associate-Data-


Practitioner
Shared by Brady on 13-05-2025
For More Free Questions and Preparation Resources

Check the Links on Last Page


Real Google Cloud Associate Data Practitioner Study Questions By Brady - Page 2

Question 1
Question Type: MultipleChoice

You have an existing weekly Storage Transfer Service transfer job from Amazon S3 to a Nearline
Cloud Storage bucket in Google Cloud. Each week, the job moves a large number of relatively
small files. As the number of files to be transferred each week has grown over time, you are at
risk of no longer completing the transfer in the allocated time frame. You need to decrease the
total transfer time by replacing the process. Your solution should minimize costs where possible.
What should you do?

Options:
A- Create a transfer job using the Google Cloud CLI, and specify the Standard storage class with
the ---custom-storage-class flag.
B- Create parallel transfer jobs using include and exclude prefixes.
C- Create a batch Dataflow job that is scheduled weekly to migrate the data from Amazon S3 to
Cloud Storage.
D- Create an agent-based transfer job that utilizes multiple transfer agents on Compute Engine
instances.

Answer:
B

Explanation:
Comprehensive and Detailed in Depth

Why B is correct:Creating parallel transfer jobs by using include and exclude prefixes allows you
to split the data into smaller chunks and transfer them in parallel.

This can significantly increase throughput and reduce the overall transfer time.

Why other options are incorrect:A: Changing the storage class to Standard will not improve
transfer speed.

C: Dataflow is a complex solution for a simple file transfer task.

D: Agent-based transfer is suitable for large files or network limitations, but not for a large
number of small files.
Real Google Cloud Associate Data Practitioner Study Questions By Brady - Page 3

Question 2
Question Type: MultipleChoice

Following a recent company acquisition, you inherited an on-premises data infrastructure that
needs to move to Google Cloud. The acquired system has 250 Apache Airflow directed acyclic
graphs (DAGs) orchestrating data pipelines. You need to migrate the pipelines to a Google Cloud
managed service with minimal effort. What should you do?

Options:
A- Convert each DAG to a Cloud Workflow and automate the execution with Cloud Scheduler.
B- Create a new Cloud Composer environment and copy DAGs to the Cloud Composer dags/
folder.
C- Create a Google Kubernetes Engine (GKE) standard cluster and deploy Airflow as a workload.
Migrate all DAGs to the new Airflow environment.
D- Create a Cloud Data Fusion instance. For each DAG, create a Cloud Data Fusion pipeline.

Answer:
B

Explanation:
Comprehensive and Detailed in Depth

Why B is correct:Cloud Composer is a managed Apache Airflow service that provides a seamless
migration path for existing Airflow DAGs.

Simply copying the DAGs to the Cloud Composer folder allows them to run directly on Google
Cloud.

Why other options are incorrect:A: Cloud Workflows is a different orchestration tool, not
compatible with Airflow DAGs.

C: GKE deployment requires setting up and managing a Kubernetes cluster, which is more
complex.

D: Cloud Data Fusion is a data integration tool, not suitable for orchestrating existing pipelines.

Cloud Composer: https://cloud.google.com/composer/docs


Real Google Cloud Associate Data Practitioner Study Questions By Brady - Page 4

Question 3
Question Type: MultipleChoice

Your company wants to implement a data transformation (ETL) pipeline for their BigQuery data
warehouse. You need to identify a managed transformation solution that allows users to develop
with SQL and JavaScript, has version control, allows for modular code, and has data quality
checks. What should you do?

Options:
A- Create a Cloud Composer environment, and orchestrate the transformations by using the
BigQueryinsertJob operator.
B- Create BigQuery scheduled queries to define the transformations in SQL.
C- Use Dataform to define the transformations in SQLX.
D- Use Dataproc to create an Apache Spark cluster and implement the transformations by using
PySpark SQL.

Answer:
C

Explanation:
Comprehensive and Detailed in Depth

Why C is correct:Dataform is a managed data transformation service that allows you to define
data pipelines using SQL and JavaScript.

It provides version control, modular code development, and data quality checks.

Why other options are incorrect:A: Cloud Composer is an orchestration tool, not a data
transformation tool.

B: Scheduled queries are not suitable for complex ETL pipelines.

D: Dataproc requires setting up a Spark cluster and writing code, which is more complex than
using Dataform.

Dataform: https://cloud.google.com/dataform/docs
Real Google Cloud Associate Data Practitioner Study Questions By Brady - Page 5

Question 4
Question Type: MultipleChoice

You created a curated dataset of market trends in BigQuery that you want to share with multiple
external partners. You want to control the rows and columns that each partner has access to. You
want to follow Google-recommended practices. What should you do?

Options:
A- Publish the dataset in Analytics Hub. Grant dataset-level access to each partner by using
subscriptions.
B- Create a separate Cloud Storage bucket for each partner. Export the dataset to each bucket
and assign each partner to their respective bucket. Grant bucket-level access by using 1AM roles.
C- Grant each partner read access to the BigQuery dataset by using 1AM roles.
D- Create a separate project for each partner and copy the dataset into each project. Publish
each dataset in Analytics Hub. Grant dataset-level access to each partner by using subscriptions.

Answer:
A

Explanation:
Comprehensive and Detailed in Depth

Why A is correct:Analytics Hub allows you to share datasets with external partners while
maintaining control over access.

Subscriptions allow granular control.

Why other options are incorrect:B: Cloud storage is for files, not bigquery datasets.

C: IAM roles do not allow for granular row and column level control.

D: Creating a separate project for each partner is complex and not scalable.

Analytics Hub: https://cloud.google.com/analytics-hub/docs

Question 5
Question Type: MultipleChoice
Real Google Cloud Associate Data Practitioner Study Questions By Brady - Page 6

You are working on a project that requires analyzing daily social media dat

a. You have 100 GB of JSON formatted data stored in Cloud Storage that keeps growing.

You need to transform and load this data into BigQuery for analysis. You want to follow the
Google-recommended approach. What should you do?

Options:
A- Manually download the data from Cloud Storage. Use a Python script to transform and upload
the data into BigQuery.
B- Use Cloud Run functions to transform and load the data into BigQuery.
C- Use Dataflow to transform the data and write the transformed data to BigQuery.
D- Use Cloud Data Fusion to transfer the data into BigQuery raw tables, and use SQL to transform
it.

Answer:
C

Explanation:
Comprehensive and Detailed in Depth

Why C is correct:Dataflow is a fully managed service for transforming and enriching data in both
batch and streaming modes.

Dataflow is googles recomended way to transform large datasets.

It is designed for parallel processing, making it suitable for large datasets.

Why other options are incorrect:A: Manual downloading and scripting is not scalable or efficient.

B: Cloud Run functions are for stateless applications, not large data transformations.

D: While Cloud Data fusion could work, Dataflow is more optimized for large scale data
transformation.

Dataflow: https://cloud.google.com/dataflow/docs

Query successful
Real Google Cloud Associate Data Practitioner Study Questions By Brady - Page 7

Question 6
Question Type: MultipleChoice

You are building a batch data pipeline to process 100 GB of structured data from multiple sources
for daily reporting. You need to transform and standardize the data prior to loading the data to
ensure that it is stored in a single dataset. You want to use a low-code solution that can be easily
built and managed. What should you do?

Options:
A- Use Cloud Data Fusion to ingest data and load the data into BigQuery. Use Looker Studio to
perform data cleaning and transformation.
B- Use Cloud Data Fusion to ingest the data, perform data cleaning and transformation, and load
the data into BigQuery.
C- Use Cloud Data Fusion to ingest the data, perform data cleaning and transformation, and load
the data into Cloud SQL for PostgreSQL.
D- Use Cloud Storage to store the data. Use Cloud Run functions to perform data cleaning and
transformation, and load the data into BigQuery.

Answer:
B

Explanation:
Comprehensive and Detailed in Depth

Why B is correct:Cloud Data Fusion is a fully managed, cloud-native data integration service for
building and managing ETL/ELT data pipelines.

It provides a graphical interface for building pipelines without coding, making it a low-code
solution.

Cloud data fusion is perfect for the ingestion, transformation and loading of data into BigQuery.

Why other options are incorrect:A: Looker studio is for visualization, not data transformation.

C: Cloud SQL is a relational database, not ideal for large-scale analytical data.

D: Cloud run is for stateless applications, not batch data processing.

Cloud Data Fusion: https://cloud.google.com/data-fusion/docs


Real Google Cloud Associate Data Practitioner Study Questions By Brady - Page 8

To Get Premium Files for Associate-Data-


Practitioner Visit
https://www.p2pexams.com/products/associate-data-p
ractitioner

For More Free Questions Visit


https://www.p2pexams.com/google/pdf/associate-data
-practitioner

You might also like