
Answer-first summary for fast verification
Answer: Sampling bias
## Detailed Explanation In this scenario, the machine learning model is exhibiting **sampling bias**, which is the correct answer (B). Here's the reasoning: ### Why Sampling Bias (B) is Correct **Sampling bias** occurs when the training dataset does not adequately represent the entire population or real-world conditions the model will encounter. In this case: - The model is disproportionately flagging individuals from a specific ethnic group. - This strongly suggests that the training data used to develop the theft detection model had an **imbalanced representation** of different ethnic groups. - The training data likely contained either: - An **overrepresentation** of that ethnic group in negative examples (e.g., footage labeled as suspicious/theft) - An **underrepresentation** of that group in normal/benign footage - Insufficient diversity across ethnic groups overall - As a result, the model learned skewed patterns that associate certain ethnic characteristics with suspicious behavior, leading to unfair and discriminatory predictions. ### Why Other Options Are Incorrect **A: Measurement Bias** - This refers to systematic errors in how data is collected or measured (e.g., faulty sensors, inconsistent labeling protocols). While measurement issues could contribute to bias, the scenario specifically describes disproportionate flagging based on ethnicity, which points more directly to representational issues in the dataset rather than measurement errors. **C: Observer Bias** - This occurs when human observers' subjective judgments influence data labeling or interpretation. Since the model is analyzing footage automatically without human intervention during inference, observer bias isn't the primary issue affecting the model's predictions. **D: Confirmation Bias** - This is a cognitive bias where people seek or interpret information to confirm existing beliefs. While confirmation bias could affect how humans design or evaluate the model, it doesn't directly describe the statistical bias in the training data that's causing the model's discriminatory behavior. ### Best Practices for Mitigation To address sampling bias in such scenarios: 1. **Collect representative training data** that reflects the diversity of the population 2. **Implement data augmentation** techniques to balance underrepresented groups 3. **Apply fairness metrics** during model evaluation to detect disparate impact 4. **Use bias mitigation techniques** such as reweighting, adversarial debiasing, or fairness constraints 5. **Continuously monitor** model performance across different demographic groups in production This type of bias is particularly critical in security and surveillance applications, as it can lead to harmful real-world consequences including discrimination and erosion of trust in AI systems.
Ultimate access to all questions.
No comments yet.
Author: LeetQuiz Editorial Team
A company uses a machine learning model to analyze security camera footage for theft detection. The model has been found to incorrectly flag individuals from a particular ethnic group at a significantly higher rate than others. What category of bias is influencing the model's predictions?
A
Measurement bias
B
Sampling bias
C
Observer bias
D
Confirmation bias