Ultimate access to all questions.
You are working on a project that involves training a TensorFlow model on a structured dataset containing 100 billion records, which are currently stored in multiple CSV files. The dataset is expected to grow over time, and you are tasked with optimizing the input/output execution performance to ensure efficient model training. Given the scale of the data, cost-effectiveness, and the need for scalability, which of the following approaches would you recommend? (Choose two options)