
Ultimate access to all questions.
Your team is working on a data pipeline that processes data from a healthcare provider. The data includes patient records with information about patient diagnoses and treatment plans. You have been tasked with ensuring the data quality of the patient records dataset. Describe the steps you would take to run data quality checks on the patient records dataset and explain how you would define data quality rules to identify and resolve data inconsistencies related to patient diagnoses and treatment plans.
A
Run data quality checks by manually inspecting each patient record and identifying inconsistencies in diagnoses and treatment plans.
B
Use AWS Glue to run data quality checks by writing custom scripts that identify inconsistencies in diagnoses and treatment plans based on specific patterns.
C
Define data quality rules using AWS Glue DataBrew by creating a new project, selecting the patient records dataset, and specifying rules to identify and resolve data inconsistencies related to patient diagnoses and treatment plans.
D
Ignore data quality checks and assume the diagnoses and treatment plans are consistent.