Ultimate access to all questions.
Upgrade Now 🚀
Sign in to unlock AI tutor
When facing performance degradation from skewed data during a join operation in Delta Lake, which method best mitigates the skew?
A
Forcing a shuffle partition increase using spark.sql.shuffle.partitions.
B
Applying salting techniques to distribute the skewed data more evenly before the join.
C
Implementing broadcast joins for all operations irrespective of data size.
D
Utilizing the OPTIMIZE command on tables post-join to improve subsequent query performance.