
Answer-first summary for fast verification
Answer: Stream data into Pub/Sub and use Dataflow to send it to Cloud Storage.
Option A is correct because Pub/Sub is designed for handling large-scale data streams with high availability and resiliency. Dataflow processes the data from Pub/Sub and sends it to Cloud Storage, which is ideal for storing structured and unstructured data in a data lake due to its durability and availability. This combination ensures efficient data processing and storage for further analysis. Options B, C, and D are incorrect as they either misuse Storage Transfer Service for data processing tasks it's not optimized for or introduce unnecessary complexities with Dataprep by Trifacta, which may not be efficient for building a data lake architecture.
Author: LeetQuiz Editorial Team
Ultimate access to all questions.
Your company produces IoT devices globally and requires a Data Lake on Google Cloud to process both structured and unstructured data from millions of these devices. Which of the following options provides a scalable and resilient architecture for this purpose on Google Cloud?
A
Stream data into Pub/Sub and use Dataflow to send it to Cloud Storage.
B
Stream data into Pub/Sub and use Storage Transfer Service to send it to BigQuery.
C
Stream data directly to Dataflow and use Dataprep by Trifacta to send it to Bigtable.
D
Stream data into Dataflow and use Storage Transfer Service to send it to BigQuery.
No comments yet.