Ultimate access to all questions.
Upgrade Now 🚀
Sign in to unlock AI tutor
In the context of Delta Lake tables experiencing high-frequency updates to existing records, which approach is most effective in reducing the performance impact on both query and ingestion operations?
A
Split the table into frequently updated and static partitions, applying updates in bulk to the dynamic partition.
B
Implement an event sourcing model, storing updates as events and materializing the final state in a separate table for queries.
C
Leverage Delta Lake‘s MERGE operation, optimizing merge conditions and clustering the table by update keys.
D
Regularly compact the Delta Lake table to merge small files and optimize the layout for faster updates and queries.