Free Professional-Data-Engineer Exam Braindumps

Pass your Google Professional Data Engineer Exam exam with these free Questions and Answers

Page 13 of 54
QUESTION 56

- (Exam Topic 6)
Your company is currently setting up data pipelines for their campaign. For all the Google Cloud Pub/Sub streaming data, one of the important business requirements is to be able to periodically identify the inputs and their timings during their campaign. Engineers have decided to use windowing and transformation in Google Cloud Dataflow for this purpose. However, when testing this feature, they find that the Cloud Dataflow job fails for the all streaming insert. What is the most likely cause of this problem?

  1. A. They have not assigned the timestamp, which causes the job to fail
  2. B. They have not set the triggers to accommodate the data coming in late, which causes the job to fail
  3. C. They have not applied a global windowing function, which causes the job to fail when the pipeline is created
  4. D. They have not applied a non-global windowing function, which causes the job to fail when the pipeline is created

Correct Answer: C

QUESTION 57

- (Exam Topic 6)
You plan to deploy Cloud SQL using MySQL. You need to ensure high availability in the event of a zone failure. What should you do?

  1. A. Create a Cloud SQL instance in one zone, and create a failover replica in another zone within the same region.
  2. B. Create a Cloud SQL instance in one zone, and create a read replica in another zone within the same region.
  3. C. Create a Cloud SQL instance in one zone, and configure an external read replica in a zone in a different region.
  4. D. Create a Cloud SQL instance in a region, and configure automatic backup to a Cloud Storage bucket in the same region.

Correct Answer: C

QUESTION 58

- (Exam Topic 6)
You’re using Bigtable for a real-time application, and you have a heavy load that is a mix of read and writes. You’ve recently identified an additional use case and need to perform hourly an analytical job to calculate certain statistics across the whole database. You need to ensure both the reliability of your production application as well as the analytical workload.
What should you do?

  1. A. Export Bigtable dump to GCS and run your analytical job on top of the exported files.
  2. B. Add a second cluster to an existing instance with a multi-cluster routing, use live-traffic app profile for your regular workload and batch-analytics profile for the analytics workload.
  3. C. Add a second cluster to an existing instance with a single-cluster routing, use live-traffic app profile for your regular workload and batch-analytics profile for the analytics workload.
  4. D. Increase the size of your existing cluster twice and execute your analytics workload on your new resized cluster.

Correct Answer: B

QUESTION 59

- (Exam Topic 6)
You are operating a streaming Cloud Dataflow pipeline. Your engineers have a new version of the pipeline with a different windowing algorithm and triggering strategy. You want to update the running pipeline with the new version. You want to ensure that no data is lost during the update. What should you do?

  1. A. Update the Cloud Dataflow pipeline inflight by passing the --update option with the --jobName set to the existing job name
  2. B. Update the Cloud Dataflow pipeline inflight by passing the --update option with the --jobName set to a new unique job name
  3. C. Stop the Cloud Dataflow pipeline with the Cancel optio
  4. D. Create a new Cloud Dataflow job with the updated code
  5. E. Stop the Cloud Dataflow pipeline with the Drain optio
  6. F. Create a new Cloud Dataflow job with the updated code

Correct Answer: A

QUESTION 60

- (Exam Topic 1)
You want to use a database of information about tissue samples to classify future tissue samples as either normal or mutated. You are evaluating an unsupervised anomaly detection method for classifying the tissue samples. Which two characteristic support this method? (Choose two.)

  1. A. There are very few occurrences of mutations relative to normal samples.
  2. B. There are roughly equal occurrences of both normal and mutated samples in the database.
  3. C. You expect future mutations to have different features from the mutated samples in the database.
  4. D. You expect future mutations to have similar features to the mutated samples in the database.
  5. E. You already have labels for which samples are mutated and which are normal in the database.

Correct Answer: BC

Page 13 of 54

Post your Comments and Discuss Google Professional-Data-Engineer exam with other Community members: