
Ultimate access to all questions.
In a scenario where you are working with a data pipeline that ingests large volumes of financial transaction data, you need to implement a partition strategy for handling the data in Azure Data Lake Storage Gen2. What partitioning approach would you recommend, and how would you implement it to ensure efficient data processing and analysis?
A
Implement a partition strategy based on the transaction amount, as this is the most important attribute for query performance.
B
Create a partition strategy based on the transaction date and time, allowing for efficient querying of data within specific time ranges.
C
Use a hash-based partitioning method to distribute the data evenly across multiple partitions, regardless of the data's characteristics.
D
Do not implement any partition strategy, as it is not necessary for financial transaction data in Azure Data Lake Storage Gen2.