
Ultimate access to all questions.
You are designing a pipeline to send application events to a Pub/Sub topic. While event ordering is not critical, you need to aggregate events across separate hourly intervals before transferring the data to BigQuery for analysis. Which technology should you use to efficiently handle and transfer this data to BigQuery, ensuring it can scale with high volumes of events?
A
Schedule a batch Dataflow job to run hourly, pulling all available messages from the Pub/Sub topic and performing the necessary aggregations.
B
Create a Cloud Function to perform the necessary data processing that executes using the Pub/Sub trigger every time a new message is published to the topic.
C
Schedule a Cloud Function to run hourly, pulling all available messages from the Pub/Sub topic and performing the necessary aggregations.
D
Create a streaming Dataflow job that reads continually from the Pub/Sub topic and performs the necessary aggregations using tumbling windows.