
Answer-first summary for fast verification
Answer: Use a Cloud Dataflow pipeline to stream data into the BigQuery table.
The correct answer is B. The requirement specifies that the data needs to be available within 1 minute of ingestion for real-time analysis, indicating a low latency requirement. Among the options, using a Cloud Dataflow pipeline to stream data into the BigQuery table (option B) is the most suitable approach to handle real-time data ingestion and subsequent analysis. Dataflow is designed for streaming data and can handle the expected growth in data volume efficiently. Batch processing methods like bq load, INSERT, and MERGE do not meet the low latency requirement for real-time analysis.
Author: LeetQuiz Editorial Team
Ultimate access to all questions.
You are responsible for managing a project that involves capturing minute-resolution data from 50,000 sensors. This data needs to be inserted into a Google BigQuery table for the purpose of conducting real-time analysis of aggregated trends. Given the substantial expected growth in data volume, it is crucial that the data be available for analysis within 1 minute of ingestion. What approach should you take to meet these requirements successfully?
A
Use bq load to load a batch of sensor data every 60 seconds.
B
Use a Cloud Dataflow pipeline to stream data into the BigQuery table.
C
Use the INSERT statement to insert a batch of data every 60 seconds.
D
Use the MERGE statement to apply updates in batch every 60 seconds.
No comments yet.