
Answer-first summary for fast verification
Answer: Switch to a larger instance type for your Dataflow workers to provide more CPU and memory resources., Increase the maximum number of workers in your pipeline to distribute the workload more evenly.
To address the bottleneck caused by high CPU utilization, **A** suggests upgrading to a larger instance type, offering more resources for processing. **C** recommends scaling out by adding more workers to distribute the load. While **B** and **E** propose using temporary buffers, these solutions introduce complexity and may not directly alleviate CPU constraints. **D**'s suggestion to change zones is unlikely to help and could worsen performance due to increased latency.
Author: LeetQuiz Editorial Team
Ultimate access to all questions.
Your Dataflow pipeline, which processes data from a Pub/Sub topic and writes it to a BigQuery dataset in the EU, is facing delays during peak periods due to all three n1-standard-1 workers reaching maximum CPU utilization. What are two effective strategies to enhance your pipeline's performance?
A
Switch to a larger instance type for your Dataflow workers to provide more CPU and memory resources.
B
Consider creating a temporary buffer table in Bigtable before transferring data to BigQuery, though this may add complexity.
C
Increase the maximum number of workers in your pipeline to distribute the workload more evenly.
D
Relocate your Dataflow pipeline to the us-central1 zone, which might not solve the issue and could increase latency.
E
Opt for creating a temporary buffer table in Cloud Spanner as an intermediate step, acknowledging potential overhead.
No comments yet.