
google - associate data practitioner pt 2

Quiz
•
Information Technology (IT)
•
Professional Development
•
Easy
bernard Echavarria
Used 2+ times
FREE Resource
Student preview

20 questions
Show all answers
1.
MULTIPLE CHOICE QUESTION
15 mins • 1 pt
You need to create a weekly aggregated sales report based on a large volume of data. You want to use Python to design an efficient process for generating this report. What should you do?
.Create a Cloud Run function that uses NumPy. Use Cloud Scheduler to schedule the function to run once a week.
.Create a Colab Enterprise notebook and use the bigframes.pandas library. Schedule the notebook to execute once a week.
Create a Cloud Data Fusion and Wrangler flow. Schedule the flow to run once a week.
.Create a Dataflow directed acyclic graph (DAG) coded in Python. Use Cloud Scheduler to schedule the code to run once a week.
2.
MULTIPLE CHOICE QUESTION
15 mins • 1 pt
Your organization has decided to move their on-premises Apache Spark-based workload to Google Cloud. You want to be able to manage the code without needing to provision and manage your own cluster. What should you do?
.Migrate the Spark jobs to Dataproc Serverless.
.Configure a Google Kubernetes Engine cluster with Spark operators, and deploy the Spark jobs
Migrate the Spark jobs to Dataproc on Google Kubernetes Engine.
.Migrate the Spark jobs to Dataproc on Compute Engine.
3.
MULTIPLE CHOICE QUESTION
15 mins • 1 pt
You are developing a data ingestion pipeline to load small CSV files into BigQuery from Cloud Storage. You want to load these files upon arrival to minimize data latency. You want to accomplish this with minimal cost and maintenance. What should you do?
.Use the bq command-line tool within a Cloud Shell instance to load the data into BigQuery
.Create a Cloud Composer pipeline to load new files from Cloud Storage to BigQuery and schedule it to run every 10 minutes.
.Create a Cloud Run function to load the data into BigQuery that is triggered when data arrives in Cloud Storage
.Create a Dataproc cluster to pull CSV files from Cloud Storage, process them using Spark, and write the results to BigQuery.
4.
MULTIPLE CHOICE QUESTION
15 mins • 1 pt
Your organization has a petabyte of application logs stored as Parquet files in Cloud Storage. You need to quickly perform a one-time SQL-based analysis of the files and join them to data that already resides in BigQuery. What should you do?
.Create a Dataproc cluster, and write a PySpark job to join the data from BigQuery to the files in Cloud Storage.
Launch a Cloud Data Fusion environment, use plugins to connect to BigQuery and Cloud Storage, and use the SQL join operation to analyze the data.
Create external tables over the files in Cloud Storage, and perform SQL joins to tables in BigQuery to analyze the data.
Use the bq load command to load the Parquet files into BigQuery, and perform SQL joins to analyze the data
5.
MULTIPLE CHOICE QUESTION
15 mins • 1 pt
Your team is building several data pipelines that contain a collection of complex tasks and dependencies that you want to execute on a schedule, in a specific order. The tasks and dependencies consist of files in Cloud Storage, Apache Spark jobs, and data in BigQuery. You need to design a system that can schedule and automate these data processing tasks using a fully managed approach. What should you do?
.Use Cloud Scheduler to schedule the jobs to run.
.Use Cloud Tasks to schedule and run the jobs asynchronously
.Create directed acyclic graphs (DAGs) in Cloud Composer. Use the appropriate operators to connect to Cloud Storage, Spark, and BigQuery
.Create directed acyclic graphs (DAGs) in Apache Airflow deployed on Google Kubernetes Engine. Use the appropriate operators to connect to Cloud Storage, Spark, and BigQuery.
6.
MULTIPLE CHOICE QUESTION
15 mins • 1 pt
You are responsible for managing Cloud Storage buckets for a research company. Your company has well-defined data tiering and retention rules. You need to optimize storage costs while achieving your data retention needs. What should you do?
.Configure the buckets to use the Archive storage class.
.Configure a lifecycle management policy on each bucket to downgrade the storage class and remove objects based on age
.Configure the buckets to use the Standard storage class and enable Object Versioning.
.Configure the buckets to use the Autoclass feature.
7.
MULTIPLE CHOICE QUESTION
15 mins • 1 pt
You are using your own data to demonstrate the capabilities of BigQuery to your organization’s leadership team. You need to perform a one- time load of the files stored on your local machine into BigQuery using as little effort as possible. What should you do?
.Write and execute a Python script using the BigQuery Storage Write API library.
Create a Dataproc cluster, copy the files to Cloud Storage, and write an Apache Spark job using the spark
.Execute the bq load command on your local machine.
.Create a Dataflow job using the Apache Beam FileIO and BigQueryIO connectors with a local runner.
Create a free account and access millions of resources
Popular Resources on Wayground
50 questions
Trivia 7/25

Quiz
•
12th Grade
11 questions
Standard Response Protocol

Quiz
•
6th - 8th Grade
11 questions
Negative Exponents

Quiz
•
7th - 8th Grade
12 questions
Exponent Expressions

Quiz
•
6th Grade
4 questions
Exit Ticket 7/29

Quiz
•
8th Grade
20 questions
Subject-Verb Agreement

Quiz
•
9th Grade
20 questions
One Step Equations All Operations

Quiz
•
6th - 7th Grade
18 questions
"A Quilt of a Country"

Quiz
•
9th Grade