Become Google Certified with updated Professional-Data-Engineer exam questions and correct answers
You are designing the architecture of your application to store data in Cloud Storage. Your application consists
of pipelines that read data from a Cloud Storage bucket that contains raw data, and write the data to a second
bucket after processing. You want to design an architecture with Cloud Storage resources that are capable of
being resilient if a Google Cloud regional failure occurs. You want to minimize the recovery point objective
(RPO) if a failure occurs, with no impact on applications that use the stored data. What should you do?
Which of these operations can you perform from the BigQuery Web UI?
You are responsible for writing your company’s ETL pipelines to run on an Apache Hadoop cluster. The pipeline will require some checkpointing and splitting pipelines. Which method should you use to write the pipelines?
You want to schedule a number of sequential load and transformation jobs Data files will be added to a Cloud
Storage bucket by an upstream process There is no fixed schedule for when the new data arrives Next, a
Dataproc job is triggered to perform some transformations and write the data to BigQuery. You then need to
run additional transformation jobs in BigQuery The transformation jobs are different for every table These
jobs might take hours to complete You need to determine the most efficient and maintainable workflow to
process hundreds of tables and provide the freshest data to your end users. What should you do?
How would you query specific partitions in a BigQuery table?
© Copyrights DumpsCertify 2025. All Rights Reserved
We use cookies to ensure your best experience. So we hope you are happy to receive all cookies on the DumpsCertify.