
Ultimate access to all questions.
You are responsible for implementing security best practices within your data pipeline system. At present, you manually execute jobs with Project Owner privileges. Your goal is to automate these tasks, which involve handling nightly batch files that contain non-public information stored in Google Cloud Storage. These files need to be processed using a Spark Scala job on a Google Cloud Dataproc cluster, and the resulting output should be deposited into Google BigQuery. What steps should you take to securely and efficiently automate this workload?
A
Restrict the Google Cloud Storage bucket so only you can see the files
B
Grant the Project Owner role to a service account, and run the job with it
C
Use a service account with the ability to read the batch files and to write to BigQuery
D
Use a user account with the Project Viewer role on the Cloud Dataproc cluster to read the batch files and write to BigQuery