
Ultimate access to all questions.
You are building an MLOps pipeline to retrain tree-based models in production. The pipeline must include components for data ingestion, data processing, model training, model evaluation, and model deployment. Your organization primarily uses PySpark for data preprocessing, and you want to minimize infrastructure management. How should you architect this pipeline?
A
Set up a TensorFlow Extended (TFX) pipeline on Vertex AI Pipelines to orchestrate the MLOps pipeline. Write a custom component for the PySpark-based workloads on Dataproc.
B
Set up a Vertex AI Pipelines to orchestrate the MLOps pipeline. Use the predefined Dataproc component for the PySpark-based workloads.
C
Set up Kubeflow Pipelines on Google Kubernetes Engine to orchestrate the MLOps pipeline. Write a custom component for the PySparkbased workloads on Dataproc.
D
Set up Cloud Composer to orchestrate the MLOps pipeline. Use Dataproc workflow templates for the PySpark-based workloads in Cloud Composer.