
Ultimate access to all questions.
You are working on a data pipeline that processes data from a marketing company. The data includes customer feedback records with information about customer satisfaction and preferences. You have been tasked with ensuring the data quality of the customer feedback records dataset. Describe the steps you would take to run data quality checks on the customer feedback records dataset and explain how you would define data quality rules to identify and filter out irrelevant or duplicate feedback records.
A
Run data quality checks by manually inspecting each customer feedback record and identifying irrelevant or duplicate records.
B
Use AWS Glue to run data quality checks by writing custom scripts that identify irrelevant or duplicate records based on specific criteria.
C
Define data quality rules using AWS Glue DataBrew by creating a new project, selecting the customer feedback records dataset, and specifying rules to identify and filter out irrelevant or duplicate records.
D
Ignore data quality checks and assume all customer feedback records are relevant and unique.