LeetQuiz Logo
Privacy Policy•contact@leetquiz.com
© 2025 LeetQuiz All rights reserved.
Google Professional Data Engineer

Google Professional Data Engineer

Get started today

Ultimate access to all questions.


After migrating a complex analytical Spark job with shuffling operations from an on-prem Hadoop cluster to Dataproc on GCS, how can you optimize its performance given the initial data is in parquet format (average size 200-400 MB each) and your organization is cost-sensitive? The job currently runs on preemptible VMs with only two non-preemptible workers.

Real Exam



Powered ByGPT-5