CertyIQ - Google - Prof Data Eng - pt 4

Quiz
•
Computers
•
University
•
Hard
Katheryne Pierce
Used 1+ times
FREE Resource
30 questions
Show all answers
1.
MULTIPLE CHOICE QUESTION
15 mins • 1 pt
You work for a bank. You have a labelled dataset that contains information on already granted loan application and whether these applications have been defaulted. You have been asked to train a model to predict default rates for credit applicants. What should you do?
Increase the size of the dataset by collecting additional data
Train a linear regression to predict a credit default risk score.
Remove the bias from the data and collect applications that have been declined loans.
Match loan applicants with their social profiles to enable feature engineering.
2.
MULTIPLE CHOICE QUESTION
15 mins • 1 pt
You need to migrate a 2TB relational database to Google Cloud Platform. You do not have the resources to significantly refactor the application that uses this database and cost to operate is of primary concern. Which service do you select for storing and serving your data?
Cloud Spanner
Cloud Bigtable
Cloud Firestore
Cloud SQL
3.
MULTIPLE CHOICE QUESTION
15 mins • 1 pt
You're using Bigtable for a real-time application, and you have a heavy load that is a mix of read and writes. You've recently identified an additional use case and need to perform hourly an analytical job to calculate certain statistics across the whole database. You need to ensure both the reliability of your production application as well as the analytical workload. What should you do?
Export Bigtable dump to GCS and run your analytical job on top of the exported files.
Add a second cluster to an existing instance with a multi-cluster routing, use live-traffic app profile for your regular workload and batch-analytics profile for the analytics workload.
Add a second cluster to an existing instance with a single-cluster routing, use live-traffic app profile for your regular workload and batch-analytics profile for the analytics workload
Increase the size of your existing cluster twice and execute your analytics workload on your new resized cluster.
4.
MULTIPLE CHOICE QUESTION
15 mins • 1 pt
You are designing an Apache Beam pipeline to enrich data from Cloud Pub/Sub with static reference data from BigQuery. The reference data is small enough to fit in memory on a single worker. The pipeline should write enriched results to BigQuery for analysis. Which job type and transforms should this pipeline use?
Batch job, PubSubIO, side-inputs
Streaming job, PubSubIO, JdbcIO, side-outputs
Streaming job, PubSubIO, BigQueryIO, side-inputs
Streaming job, PubSubIO, BigQueryIO, side-outputs
5.
MULTIPLE SELECT QUESTION
15 mins • 1 pt
You have a data pipeline that writes data to Cloud Bigtable using well-designed row keys. You want to monitor your pipeline to determine when to increase the size of your Cloud Bigtable cluster. Which two actions can you take to accomplish this? (Choose two.)
Review Key Visualizer metrics. Increase the size of the Cloud Bigtable cluster when the Read pressure index is above 100.
Review Key Visualizer metrics. Increase the size of the Cloud Bigtable cluster when the Write pressure index is above 100.
Monitor the latency of write operations. Increase the size of the Cloud Bigtable cluster when there is a sustained increase in write latency.
Monitor storage utilization. Increase the size of the Cloud Bigtable cluster when utilization increases above 70% of max capacity.
Monitor latency of read operations. Increase the size of the Cloud Bigtable cluster of read operations take longer than 100 ms.
6.
MULTIPLE CHOICE QUESTION
15 mins • 1 pt
Store the social media posts and the data extracted from the API in BigQuery.
Store the social media posts and the data extracted from the API in Cloud SQL.
Store the raw social media posts in Cloud Storage, and write the data extracted from the API into BigQuery
Feed to social media posts into the API directly from the source, and write the extracted data from the API into BigQuery.
7.
MULTIPLE CHOICE QUESTION
15 mins • 1 pt
You store historic data in Cloud Storage. You need to perform analytics on the historic data. You want to use a solution to detect invalid data entries and perform data transformations that will not require programming or knowledge of SQL. What should you do?
Use Cloud Dataflow with Beam to detect errors and perform transformations.
Use Cloud Dataprep with recipes to detect errors and perform transformations.
Use Cloud Dataproc with a Hadoop job to detect errors and perform transformations.
Use federated tables in BigQuery with queries to detect errors and perform transformations.
Create a free account and access millions of resources
Similar Resources on Wayground
25 questions
DBC-IT105-Prelim Quiz

Quiz
•
University
35 questions
ITD11303 DATABASE

Quiz
•
University
25 questions
BASIC PROGRAMMING

Quiz
•
University
26 questions
Week 4 - RELATIONAL DATA MODEL

Quiz
•
University
25 questions
GDSC FET Jain University Study Jam 106

Quiz
•
University
35 questions
AZ-400 Quiz

Quiz
•
University
30 questions
Cloud Computing - Bangkit Academy 2024 H1

Quiz
•
University
30 questions
Quiz TBD Pra UAS

Quiz
•
University
Popular Resources on Wayground
18 questions
Writing Launch Day 1

Lesson
•
3rd Grade
11 questions
Hallway & Bathroom Expectations

Quiz
•
6th - 8th Grade
11 questions
Standard Response Protocol

Quiz
•
6th - 8th Grade
40 questions
Algebra Review Topics

Quiz
•
9th - 12th Grade
4 questions
Exit Ticket 7/29

Quiz
•
8th Grade
10 questions
Lab Safety Procedures and Guidelines

Interactive video
•
6th - 10th Grade
19 questions
Handbook Overview

Lesson
•
9th - 12th Grade
20 questions
Subject-Verb Agreement

Quiz
•
9th Grade