
Answer-first summary for fast verification
Answer: Fairness
## Analysis of the Scenario The scenario describes an AI-powered resume screening system trained on a dataset that "did not represent all demographics." This indicates the training data was imbalanced or biased, potentially leading the model to perform poorly or unfairly for underrepresented demographic groups. ## Core Dimensions of Responsible AI AWS and industry frameworks typically identify several core dimensions of responsible AI: - **Fairness**: Ensuring AI systems treat all individuals and groups equitably, without bias or discrimination. - **Explainability**: Making AI decisions understandable to humans. - **Privacy and Security**: Protecting data and ensuring secure AI operations. - **Transparency**: Being open about how AI systems are developed, deployed, and used. ## Why Fairness (Option A) is the Correct Answer 1. **Direct Connection to Data Bias**: The scenario explicitly mentions non-representative data across demographics. This is a classic fairness issue where biased training data can lead to discriminatory outcomes against certain groups. 2. **Impact on Decision-Making**: In resume screening, fairness is critical because biased models could systematically disadvantage candidates from underrepresented demographics, perpetuating existing inequalities in hiring. 3. **AWS Responsible AI Framework**: AWS emphasizes fairness as a key pillar, specifically addressing issues of biased training data and ensuring AI systems don't discriminate against protected groups. ## Why Other Options Are Less Suitable - **Explainability (B)**: While important, the scenario doesn't mention issues with understanding how the model makes decisions. The problem is rooted in the input data, not the model's interpretability. - **Privacy and Security (C)**: The scenario doesn't describe data breaches, unauthorized access, or privacy violations. The issue is about data composition, not data protection. - **Transparency (D)**: Although transparency about data limitations would be valuable, the core issue is the biased outcomes that may result from the data imbalance, which falls squarely under fairness concerns. ## Best Practices Consideration In responsible AI development, addressing fairness involves: 1. Using diverse, representative training data 2. Testing for disparate impact across demographic groups 3. Implementing bias detection and mitigation techniques 4. Continuously monitoring for fairness in production The scenario highlights a failure in the first step - using non-representative data - making fairness the most directly relevant dimension.
Author: LeetQuiz Editorial Team
Ultimate access to all questions.
No comments yet.
A company developed an AI-powered resume screening system trained on a large dataset. The dataset consisted of resumes that did not represent all demographics.
Which core dimension of responsible AI does this scenario illustrate?
A
Fairness
B
Explainability
C
Privacy and security
D
Transparency