CertyIQ - Google - Prof Data Eng - pt 7

CertyIQ - Google - Prof Data Eng - pt 7

University

30 Qs

quiz-placeholder

Similar activities

The GCCP Quiz

The GCCP Quiz

University

25 Qs

Quiz on Aggregate Functions in SQL

Quiz on Aggregate Functions in SQL

University

25 Qs

Database Questions

Database Questions

KG - Professional Development

31 Qs

Google Prof Cloud Archi - pt 3

Google Prof Cloud Archi - pt 3

University

30 Qs

CertyIQ - Google - Prof Data Eng - pt 9

CertyIQ - Google - Prof Data Eng - pt 9

University

30 Qs

Practice Questions DP 300

Practice Questions DP 300

University

25 Qs

Microsoft Access Quiz

Microsoft Access Quiz

University

25 Qs

Value Added Course_NoSQL Databases

Value Added Course_NoSQL Databases

University

25 Qs

CertyIQ - Google - Prof Data Eng - pt 7

CertyIQ - Google - Prof Data Eng - pt 7

Assessment

Quiz

Computers

University

Medium

Created by

Katheryne Pierce

Used 6+ times

FREE Resource

30 questions

Show all answers

1.

MULTIPLE CHOICE QUESTION

15 mins • 1 pt

You need to give new website users a globally unique identifier (GUID) using a service that takes in data points and returns a GUID. This data is sourced from both internal and external systems via HTTP calls that you will make via microservices within your pipeline. There will be tens of thousands of messages per second and that can be multithreaded. and you worry about the backpressure on the system. How should you design your pipeline to minimize that backpressure?

Call out to the service via HTTP.

Create the pipeline statically in the class definition

Create a new object in the startBundle method of DoFn.

Batch the job into ten-second increments.

2.

MULTIPLE CHOICE QUESTION

15 mins • 1 pt

You are migrating your data warehouse to Google Cloud and decommissioning your on-premises data center. Because this is a priority for your company, you know that bandwidth will be made available for the initial data load to the cloud. The files being transferred are not large in number, but each file is 90 GB. Additionally, you want your transactional systems to continually update the warehouse on Google Cloud in real time. What tools should you use to migrate the data and ensure that it continues to write to your warehouse?

Storage Transfer Service for the migration; Pub/Sub and Cloud Data Fusion for the real-time updates

BigQuery Data Transfer Service for the migration; Pub/Sub and Dataproc for the real-time updates

gsutil for the migration; Pub/Sub and Dataflow for the real-time updates

gsutil for both the migration and the real-time updates

3.

MULTIPLE CHOICE QUESTION

15 mins • 1 pt

You are using Bigtable to persist and serve stock market data for each of the major indices. To serve the trading application, you need to access only the most recent stock prices that are streaming in. How should you design your row key and tables to ensure that you can access the data with the simplest query?

Create one unique table for all of the indices, and then use the index and timestamp as the row key design

Create one unique table for all of the indices, and then use a reverse timestamp as the row key design.

For each index, have a separate table and use a timestamp as the row key design.

For each index, have a separate table and use a reverse timestamp as the row key design.

4.

MULTIPLE CHOICE QUESTION

15 mins • 1 pt

You are building a report-only data warehouse where the data is streamed into BigQuery via the streaming API. Following Google's best practices, you have both a staging and a production table for the data. How should you design your data loading to ensure that there is only one master dataset without affecting performance on either the ingestion or reporting pieces?

Have a staging table that is an append-only model, and then update the production table every three hours with the changes written to staging.

Have a staging table that is an append-only model, and then update the production table every ninety minutes with the changes written to staging.

Have a staging table that moves the staged data over to the production table and deletes the contents of the staging table every three hours.

Have a staging table that moves the staged data over to the production table and deletes the contents of the staging table every thirty minutes.

5.

MULTIPLE CHOICE QUESTION

15 mins • 1 pt

You issue a new batch job to Dataflow. The job starts successfully, processes a few elements, and then suddenly fails and shuts down. You navigate to the Dataflow monitoring interface where you find errors related to a particular DoFn in your pipeline. What is the most likely cause of the errors?

Job validation

Exceptions in worker code

Graph or pipeline construction

Insufficient permissions

6.

MULTIPLE CHOICE QUESTION

15 mins • 1 pt

Your new customer has requested daily reports that show their net consumption of Google Cloud compute resources and who used the resources. You need to quickly and efficiently generate these daily reports. What should you do?

Do daily exports of Cloud Logging data to BigQuery. Create views filtering by project, log type, resource, and user

Filter data in Cloud Logging by project, resource, and user; then export the data in CSV format.

Filter data in Cloud Logging by project, log type, resource, and user, then import the data into BigQuery.

Export Cloud Logging data to Cloud Storage in CSV format. Cleanse the data using Dataprep, filtering by project, resource, and user.

7.

MULTIPLE CHOICE QUESTION

15 mins • 1 pt

Media Image

The Development and External teams have the project viewer Identity and Access Management (IAM) role in a folder named Visualization. You want the Development Team to be able to read data from both Cloud Storage and BigQuery, but the External Team should only be able to read data from BigQuery. What should you do?

Remove Cloud Storage IAM permissions to the External Team on the acme-raw-data project.

Create Virtual Private Cloud (VPC) firewall rules on the acme-raw-data project that deny all ingress traffic from the External Team CIDR range.

Create a VPC Service Controls perimeter containing both projects and BigQuery as a restricted API. Add the External Team users to the perimeter's Access Level.

Create a VPC Service Controls perimeter containing both projects and Cloud Storage as a restricted API. Add the Development Team users to the perimeter's Access Level.

Create a free account and access millions of resources

Create resources
Host any resource
Get auto-graded reports
or continue with
Microsoft
Apple
Others
By signing up, you agree to our Terms of Service & Privacy Policy
Already have an account?