
Ultimate access to all questions.
Your team builds services on Google Cloud. You need to process messages from a Pub/Sub topic and store them, ensuring each message is processed exactly once to prevent data duplication or conflicts. The solution must be the simplest and most cost-effective. What is the recommended approach?
A
Process the messages with a Dataproc job, and write the output to storage.
B
Process the messages with a Dataflow streaming pipeline using Apache Beam's PubSubIO package, and write the output to storage.
C
Process the messages with a Cloud Function, and write the results to a BigQuery location where you can run a job to deduplicate the data.
D
Retrieve the messages with a Dataflow streaming pipeline, store them in Cloud Bigtable, and use another Dataflow streaming pipeline to deduplicate messages.