Consider a scenario where you have a large dataset in Apache Spark and you need to apply a pre-trained machine learning model to each row in parallel. You are advised to use Pandas UDFs for this task. Why is Apache Arrow considered crucial in this context?