
Ultimate access to all questions.
As a Microsoft Fabric Analytics Engineer Associate, you are designing a data pipeline to ingest data from multiple sources, including a high-volume social media platform known for its varying data quality, into a data warehouse. The project has strict compliance requirements and must ensure data integrity without exceeding the allocated budget. Considering these constraints, which approach would you take to address the data quality issues effectively? (Choose one option.)
A
Implement strict data validation rules at the social media platform's API level to prevent low-quality data from entering the pipeline, assuming you have administrative access to the platform.
B
Integrate a data cleansing and transformation layer within the pipeline that automatically identifies, corrects, or flags data quality issues before loading the data into the warehouse.
C
Exclude the social media data entirely from the pipeline to avoid the risk of compromising the data warehouse's integrity, despite the potential loss of insights.
D
Create a parallel processing pipeline for social media data with separate quality controls and merge it with the main data warehouse only after ensuring compliance and quality standards are met.