
Ultimate access to all questions.
You are managing a demand forecasting pipeline in production. This pipeline leverages Dataflow to preprocess raw data before model training and prediction. The preprocessing step involves applying Z-score normalization to data stored in BigQuery, and the normalized data is then written back to BigQuery. New training data is added to the pipeline on a weekly basis. To enhance the efficiency of this process and reduce both computation time and manual intervention, what should you do?
A
Normalize the data using Google Kubernetes Engine.
B
Translate the normalization algorithm into SQL for use with BigQuery.
C
Use the normalizer_fn argument in TensorFlow's Feature Column API._
D
Normalize the data with Apache Spark using the Dataproc connector for BigQuery.