
Answer-first summary for fast verification
Answer: 1. Using Vertex AI Pipelines, add a component to divide the data into training and evaluation sets, and add another component for feature engineering. 2. Enable autologging of metrics in the training component. 3. Compare pipeline runs in Vertex AI Experiments.
Option A is the correct answer because it leverages Vertex AI Pipelines and Vertex AI Experiments, providing a flexible and controlled environment for data splitting, feature engineering, and model training. Vertex AI autologging simplifies capturing evaluation metrics during training, and Vertex AI Experiments offer a centralized interface to compare metrics across different pipeline runs. In comparison, options C and D, which use BigQuery ML, provide less control over feature engineering and model comparisons. Option B, while also using Vertex AI Pipelines, recommends using Vertex ML Metadata for model comparison, which is less comprehensive than Vertex AI Experiments.
Author: LeetQuiz Editorial Team
Ultimate access to all questions.
No comments yet.
You are tasked with developing a machine learning pipeline for training an XGBoost classification model using tabular data stored in a BigQuery table. Your goal is to ensure that the pipeline effectively handles data splitting, feature engineering, and model evaluation, and allows for easy comparison of different models. The required steps are: 1. Randomly split the data into training and evaluation datasets in a 65/35 ratio. 2. Conduct feature engineering to prepare the data for training. 3. Obtain evaluation metrics for the model's performance. 4. Compare the performance of models trained in different pipeline executions. Which approach should you take to achieve these requirements?
A
B
C
D