You work for a bank. You have a labelled dataset that contains information on already granted loan application and whether these applications have been defaulted. You have been asked to train a model to predict default rates for credit applicants. What should you do?
CertyIQ - Google - Prof Data Eng - pt 4

Quiz
•
Computers
•
University
•
Hard
Katheryne Pierce
Used 1+ times
FREE Resource
30 questions
Show all answers
1.
MULTIPLE CHOICE QUESTION
15 mins • 1 pt
Increase the size of the dataset by collecting additional data
Train a linear regression to predict a credit default risk score.
Remove the bias from the data and collect applications that have been declined loans.
Match loan applicants with their social profiles to enable feature engineering.
2.
MULTIPLE CHOICE QUESTION
15 mins • 1 pt
You need to migrate a 2TB relational database to Google Cloud Platform. You do not have the resources to significantly refactor the application that uses this database and cost to operate is of primary concern. Which service do you select for storing and serving your data?
Cloud Spanner
Cloud Bigtable
Cloud Firestore
Cloud SQL
3.
MULTIPLE CHOICE QUESTION
15 mins • 1 pt
You're using Bigtable for a real-time application, and you have a heavy load that is a mix of read and writes. You've recently identified an additional use case and need to perform hourly an analytical job to calculate certain statistics across the whole database. You need to ensure both the reliability of your production application as well as the analytical workload. What should you do?
Export Bigtable dump to GCS and run your analytical job on top of the exported files.
Add a second cluster to an existing instance with a multi-cluster routing, use live-traffic app profile for your regular workload and batch-analytics profile for the analytics workload.
Add a second cluster to an existing instance with a single-cluster routing, use live-traffic app profile for your regular workload and batch-analytics profile for the analytics workload
Increase the size of your existing cluster twice and execute your analytics workload on your new resized cluster.
4.
MULTIPLE CHOICE QUESTION
15 mins • 1 pt
You are designing an Apache Beam pipeline to enrich data from Cloud Pub/Sub with static reference data from BigQuery. The reference data is small enough to fit in memory on a single worker. The pipeline should write enriched results to BigQuery for analysis. Which job type and transforms should this pipeline use?
Batch job, PubSubIO, side-inputs
Streaming job, PubSubIO, JdbcIO, side-outputs
Streaming job, PubSubIO, BigQueryIO, side-inputs
Streaming job, PubSubIO, BigQueryIO, side-outputs
5.
MULTIPLE SELECT QUESTION
15 mins • 1 pt
You have a data pipeline that writes data to Cloud Bigtable using well-designed row keys. You want to monitor your pipeline to determine when to increase the size of your Cloud Bigtable cluster. Which two actions can you take to accomplish this? (Choose two.)
Review Key Visualizer metrics. Increase the size of the Cloud Bigtable cluster when the Read pressure index is above 100.
Review Key Visualizer metrics. Increase the size of the Cloud Bigtable cluster when the Write pressure index is above 100.
Monitor the latency of write operations. Increase the size of the Cloud Bigtable cluster when there is a sustained increase in write latency.
Monitor storage utilization. Increase the size of the Cloud Bigtable cluster when utilization increases above 70% of max capacity.
Monitor latency of read operations. Increase the size of the Cloud Bigtable cluster of read operations take longer than 100 ms.
6.
MULTIPLE CHOICE QUESTION
15 mins • 1 pt
Store the social media posts and the data extracted from the API in BigQuery.
Store the social media posts and the data extracted from the API in Cloud SQL.
Store the raw social media posts in Cloud Storage, and write the data extracted from the API into BigQuery
Feed to social media posts into the API directly from the source, and write the extracted data from the API into BigQuery.
7.
MULTIPLE CHOICE QUESTION
15 mins • 1 pt
You store historic data in Cloud Storage. You need to perform analytics on the historic data. You want to use a solution to detect invalid data entries and perform data transformations that will not require programming or knowledge of SQL. What should you do?
Use Cloud Dataflow with Beam to detect errors and perform transformations.
Use Cloud Dataprep with recipes to detect errors and perform transformations.
Use Cloud Dataproc with a Hadoop job to detect errors and perform transformations.
Use federated tables in BigQuery with queries to detect errors and perform transformations.
Create a free account and access millions of resources
Similar Resources on Quizizz
30 questions
Google Prof Cloud Archi - pt 3

Quiz
•
University
30 questions
Google Prof Cloud Archi - pt 4

Quiz
•
University
25 questions
Practice Questions DP 300

Quiz
•
University
25 questions
Value Added Course_NoSQL Databases

Quiz
•
University
30 questions
Information Management (LEC)

Quiz
•
University
30 questions
Google Prof Cloud Archi - pt 9

Quiz
•
University
25 questions
Cloud Infrastructure Mechanisms

Quiz
•
University
30 questions
Google Prof Cloud Archi - pt 5

Quiz
•
University
Popular Resources on Quizizz
15 questions
Character Analysis

Quiz
•
4th Grade
17 questions
Chapter 12 - Doing the Right Thing

Quiz
•
9th - 12th Grade
10 questions
American Flag

Quiz
•
1st - 2nd Grade
20 questions
Reading Comprehension

Quiz
•
5th Grade
30 questions
Linear Inequalities

Quiz
•
9th - 12th Grade
20 questions
Types of Credit

Quiz
•
9th - 12th Grade
18 questions
Full S.T.E.A.M. Ahead Summer Academy Pre-Test 24-25

Quiz
•
5th Grade
14 questions
Misplaced and Dangling Modifiers

Quiz
•
6th - 8th Grade