CertyIQ - Google - Prof Data Eng - pt 4

CertyIQ - Google - Prof Data Eng - pt 4

University

30 Qs

quiz-placeholder

Similar activities

Practice Questions DP 300

Practice Questions DP 300

University

25 Qs

ITD11303 DATABASE

ITD11303 DATABASE

University

35 Qs

Google Prof Cloud Archi - pt 5

Google Prof Cloud Archi - pt 5

University

30 Qs

Database Questions

Database Questions

KG - Professional Development

31 Qs

DBC-IT105-Prelim Quiz

DBC-IT105-Prelim Quiz

University

25 Qs

Google Prof Cloud Archi - pt 3

Google Prof Cloud Archi - pt 3

University

30 Qs

Google Prof Cloud Archi - pt 4

Google Prof Cloud Archi - pt 4

University

30 Qs

subquery

subquery

University

26 Qs

CertyIQ - Google - Prof Data Eng - pt 4

CertyIQ - Google - Prof Data Eng - pt 4

Assessment

Quiz

Computers

University

Hard

Created by

Katheryne Pierce

Used 1+ times

FREE Resource

30 questions

Show all answers

1.

MULTIPLE CHOICE QUESTION

15 mins • 1 pt

You work for a bank. You have a labelled dataset that contains information on already granted loan application and whether these applications have been defaulted. You have been asked to train a model to predict default rates for credit applicants. What should you do?

Increase the size of the dataset by collecting additional data

Train a linear regression to predict a credit default risk score.

Remove the bias from the data and collect applications that have been declined loans.

Match loan applicants with their social profiles to enable feature engineering.

2.

MULTIPLE CHOICE QUESTION

15 mins • 1 pt

You need to migrate a 2TB relational database to Google Cloud Platform. You do not have the resources to significantly refactor the application that uses this database and cost to operate is of primary concern. Which service do you select for storing and serving your data?

Cloud Spanner

Cloud Bigtable

Cloud Firestore

Cloud SQL

3.

MULTIPLE CHOICE QUESTION

15 mins • 1 pt

You're using Bigtable for a real-time application, and you have a heavy load that is a mix of read and writes. You've recently identified an additional use case and need to perform hourly an analytical job to calculate certain statistics across the whole database. You need to ensure both the reliability of your production application as well as the analytical workload. What should you do?

Export Bigtable dump to GCS and run your analytical job on top of the exported files.

Add a second cluster to an existing instance with a multi-cluster routing, use live-traffic app profile for your regular workload and batch-analytics profile for the analytics workload.

Add a second cluster to an existing instance with a single-cluster routing, use live-traffic app profile for your regular workload and batch-analytics profile for the analytics workload

Increase the size of your existing cluster twice and execute your analytics workload on your new resized cluster.

4.

MULTIPLE CHOICE QUESTION

15 mins • 1 pt

You are designing an Apache Beam pipeline to enrich data from Cloud Pub/Sub with static reference data from BigQuery. The reference data is small enough to fit in memory on a single worker. The pipeline should write enriched results to BigQuery for analysis. Which job type and transforms should this pipeline use?

Batch job, PubSubIO, side-inputs

Streaming job, PubSubIO, JdbcIO, side-outputs

Streaming job, PubSubIO, BigQueryIO, side-inputs

Streaming job, PubSubIO, BigQueryIO, side-outputs

5.

MULTIPLE SELECT QUESTION

15 mins • 1 pt

You have a data pipeline that writes data to Cloud Bigtable using well-designed row keys. You want to monitor your pipeline to determine when to increase the size of your Cloud Bigtable cluster. Which two actions can you take to accomplish this? (Choose two.)

Review Key Visualizer metrics. Increase the size of the Cloud Bigtable cluster when the Read pressure index is above 100.

Review Key Visualizer metrics. Increase the size of the Cloud Bigtable cluster when the Write pressure index is above 100.

Monitor the latency of write operations. Increase the size of the Cloud Bigtable cluster when there is a sustained increase in write latency.

Monitor storage utilization. Increase the size of the Cloud Bigtable cluster when utilization increases above 70% of max capacity.

Monitor latency of read operations. Increase the size of the Cloud Bigtable cluster of read operations take longer than 100 ms.

6.

MULTIPLE CHOICE QUESTION

15 mins • 1 pt

Media Image

Store the social media posts and the data extracted from the API in BigQuery.

Store the social media posts and the data extracted from the API in Cloud SQL.

Store the raw social media posts in Cloud Storage, and write the data extracted from the API into BigQuery

Feed to social media posts into the API directly from the source, and write the extracted data from the API into BigQuery.

7.

MULTIPLE CHOICE QUESTION

15 mins • 1 pt

You store historic data in Cloud Storage. You need to perform analytics on the historic data. You want to use a solution to detect invalid data entries and perform data transformations that will not require programming or knowledge of SQL. What should you do?

Use Cloud Dataflow with Beam to detect errors and perform transformations.

Use Cloud Dataprep with recipes to detect errors and perform transformations.

Use Cloud Dataproc with a Hadoop job to detect errors and perform transformations.

Use federated tables in BigQuery with queries to detect errors and perform transformations.

Create a free account and access millions of resources

Create resources
Host any resource
Get auto-graded reports
or continue with
Microsoft
Apple
Others
By signing up, you agree to our Terms of Service & Privacy Policy
Already have an account?