
Ultimate access to all questions.
You recently trained a XGBoost model for predicting customer churn that you plan to deploy to a production environment for real-time inference. Before sending a predict request to your model's binary, you need to perform a data preprocessing step to clean and structure the incoming data format. This preprocessing service needs to expose a REST API that accepts requests within your secure internal VPC Service Controls and returns the processed data to the model. You aim to configure this preprocessing step while minimizing both cost and the complexity of deployment. What approach should you take?
A
Store a pickled model in Cloud Storage. Build a Flask-based app, package the app in a custom container image, and deploy the model to Vertex AI Endpoints.
B
Build a Flask-based app, package the app and a pickled model in a custom container image, and deploy the model to Vertex AI Endpoints.
C
Build a custom predictor class based on XGBoost Predictor from the Vertex AI SDK, package it and a pickled model in a custom container image based on a Vertex built-in image, and deploy the model to Vertex AI Endpoints.
D
Build a custom predictor class based on XGBoost Predictor from the Vertex AI SDK, and package the handler in a custom container image based on a Vertex built-in container image. Store a pickled model in Cloud Storage, and deploy the model to Vertex AI Endpoints.