
Ultimate access to all questions.
You have trained a machine learning model using XGBoost in Python intended for online serving. The model prediction service is expected to be invoked by a backend service developed in Golang, which operates on a Google Kubernetes Engine (GKE) cluster. The ML model necessitates certain preprocessing and postprocessing steps to function correctly at serving time. Your objectives include minimizing code changes and infrastructure maintenance, and deploying the model into production swiftly. Given these requirements, what should you do to implement the preprocessing and postprocessing steps and ensure efficient deployment?
A
Use FastAPI to implement an HTTP server. Create a Docker image that runs your HTTP server, and deploy it on your organization’s GKE cluster.
B
Use FastAPI to implement an HTTP server. Create a Docker image that runs your HTTP server, Upload the image to Vertex AI Model Registry and deploy it to a Vertex AI endpoint.
C
Use the Predictor interface to implement a custom prediction routine. Build the custom container, upload the container to Vertex AI Model Registry and deploy it to a Vertex AI endpoint.
D
Use the XGBoost prebuilt serving container when importing the trained model into Vertex AI. Deploy the model to a Vertex AI endpoint. Work with the backend engineers to implement the pre- and postprocessing steps in the Golang backend service.