
Ultimate access to all questions.
When integrating Azure Databricks with multiple external data sources, which strategy best ensures the integrity and reliability of data ingestion pipelines during integration testing, considering the variability in external source availability and data formats?
A
Implementing Azure Event Hubs to simulate data flow from external sources, enabling controlled load testing and format variability
B
Mocking external data sources within Databricks notebooks for testing pipeline integrity without actual data ingestion
C
Utilizing a staging environment that mirrors production with live connections to external sources, employing throttling to simulate variable data loads
D
Conducting manual tests by intermittently disabling connections to external sources to understand failure points