
Ultimate access to all questions.
You are part of a data science team at a bank tasked with creating a machine learning model to predict loan default risk. You have gathered and meticulously cleaned a vast amount of training data, consisting of hundreds of millions of records, which are now stored in a BigQuery table. Your next step is to develop and compare multiple models on this extensive dataset using TensorFlow and Vertex AI. Given the scale of the data, you want to ensure the data ingestion phase is efficient and scalable, minimizing any potential bottlenecks. What should you do to achieve this?
A
Use the BigQuery client library to load data into a dataframe, and use tf.data.Dataset.from_tensor_slices() to read it.
B
Export data to CSV files in Cloud Storage, and use tf.data.TextLineDataset() to read them.
C
Convert the data into TFRecords, and use tf.data.TFRecordDataset() to read them.
D
Use TensorFlow I/O’s BigQuery Reader to directly read the data.