Ultimate access to all questions.
Upgrade Now 🚀
Sign in to unlock AI tutor
When processing a large DataFrame in Spark with multiple transformations, which method ensures the most efficient execution?
A
Use UDFs for all transformations to improve readability and maintainability.
B
Chain transformations together before executing any actions to leverage Spark‘s lazy evaluation.
C
Apply transformations sequentially, and call .cache() after each transformation.
D
Execute an action after each transformation to immediately view results and ensure correctness.