
Ultimate access to all questions.
Ensuring data consistency across development, staging, and production environments is crucial for CI/CD pipelines involving data transformation jobs in Azure Databricks. Which advanced approach would you adopt to automate data consistency validation within your deployment process?
A
Writing custom Spark jobs in Databricks to compare dataset snapshots across environments, triggered as part of the deployment pipeline.
B
Utilizing Azure DevOps to run data validation scripts post-deployment, with manual interventions required to address inconsistencies.
C
Leveraging Azure Data Factory to orchestrate data movement and validation jobs between environments, using checksums for consistency checks.
D
Implementing a blockchain-based ledger within Databricks to record and verify data consistency across environments automatically.