
Ultimate access to all questions.
Deep dive into the quiz with AI chat providers.
We prepare a focused prompt with your quiz and certificate details so each AI can offer a more tailored, in-depth explanation.
A candidate wishes to regress the earnings of individuals. He uses a dummy variable "Men" which takes on the value "1" for men and is "0" otherwise and a second dummy variable "Women" which similarly takes on the value "1" for women and is "0" otherwise. If it's known that men typically earn more than women, we would expect:
A
A statistically significant difference in means
B
None of the OLS estimators to exist as men's and women's earnings would exhibit perfect multicollinearity
C
The slope coefficients of men to be equal to that of women
D
The coefficient of men to be positive and that of women to be negative
Explanation:
Perfect multicollinearity occurs when there is an exact linear relationship between independent variables in a regression model. In this case:
The candidate creates two dummy variables:
For every observation in the dataset:
This creates a perfect linear relationship: Men + Women = 1 for every observation
In matrix algebra terms, this means the design matrix X has linearly dependent columns, making X'X singular (non-invertible).
Ordinary Least Squares (OLS) requires the inversion of X'X to estimate coefficients. When perfect multicollinearity exists, this matrix cannot be inverted, making OLS estimators impossible to compute.
Why other options are incorrect:
Proper approach: To avoid perfect multicollinearity when using dummy variables for categorical data, one should use k-1 dummy variables for k categories (e.g., include only a "Men" dummy variable, with "Women" as the reference category).