
Answer-first summary for fast verification
Answer: Implementing Azure Data Factory to orchestrate and monitor data movements, with consistency checks run as Databricks jobs
For geo-distributed teams using Azure Databricks, ensuring data consistency and synchronization involves both orchestration and validation of data pipelines. Azure Data Factory (ADF) excels at: - Orchestrating complex ETL workflows across multiple locations and systems, - Monitoring data pipelines in real time with alerting and logging, - Running consistency checks using Databricks notebooks or jobs as part of the pipeline steps. This approach is scalable, automated, and integrates well with other Azure services. Why the other options are less correct: - **A**: Periodic validation jobs and logging are useful, but lack orchestration of actual data movement and don’t ensure synchronization by themselves. - **B**: Delta Lake and Unity Catalog provide data reliability and governance, but not orchestration or cross-region consistency checks. They’re part of the solution but not the primary mechanism for monitoring and synchronization. - **D**: Manual coordination and emails are prone to human error and not a scalable or reliable strategy.
Author: LeetQuiz Editorial Team
Ultimate access to all questions.
No comments yet.
How can you ensure data consistency and synchronization across geo-distributed teams using Azure Databricks?
A
Configuring periodic data consistency validation jobs within Databricks, with results logged to Azure Log Analytics for monitoring
B
Using Azure Databricks Delta Lake for ACID transactions and leveraging Unity Catalog for data governance across locations
C
Implementing Azure Data Factory to orchestrate and monitor data movements, with consistency checks run as Databricks jobs
D
Manually coordinating data updates across teams, relying on email notifications for data changes