QM 2 Flashcards
(61 cards)
What is conditional heteroskedasticity?
It occurs when the error variance in a regression is correlated with or conditional on the values of the independent variables.
What are the main consequences of conditional heteroskedasticity for statistical inference?
Standard errors of regression coefficients are underestimated t-statistics are inflated and Type I errors rejecting a true null hypothesis are more likely. The F-test for overall significance also becomes unreliable.
What is the purpose of the Breusch-Pagan (BP) test?
To test for conditional heteroskedasticity in a regression model.
What is the null hypothesis of the Breusch-Pagan (BP) test?
There is no conditional heteroskedasticity meaning the regression’s squared residuals are uncorrelated with the independent variables.
How is the Breusch-Pagan (BP) test statistic calculated and what is its distribution?
The BP test statistic is n multiplied by R-squared from a regression of the squared residuals on the original independent variables. It is approximately chi-square distributed with k degrees of freedom where k is the number of independent variables.
How do you correct for conditional heteroskedasticity?
By computing robust standard errors also known as heteroskedasticity-consistent or White-corrected standard errors.
What is serial correlation or autocorrelation in a regression model?
It occurs when regression errors are correlated across observations typically in time-series regressions.
What are the consequences of serial correlation if one of the independent variables is a lagged value of the dependent variable?
All parameter estimates (coefficients) become inconsistent and invalid. Standard errors are also invalid.
What are the consequences of positive serial correlation if no independent variable is a lagged dependent variable?
Coefficient estimates are consistent but standard errors are typically underestimated leading to inflated t-statistics and more Type I errors. The F-statistic may also be inflated.
What is the Breusch-Godfrey (BG) test used for?
To detect serial correlation of a pre-designated order p in regression residuals.
What is the null hypothesis of the Breusch-Godfrey (BG) test?
There is no serial correlation in the model’s residuals up to the specified lag p.
How are robust standard errors (like Newey-West) useful in the context of serial correlation?
They adjust the coefficient standard errors to account for serial correlation and also correct for conditional heteroskedasticity allowing for more reliable statistical inference.
What is multicollinearity in a multiple regression model?
It occurs when two or more independent variables are highly correlated or when there is an approximate linear relationship among independent variables.
What are the primary consequences of multicollinearity?
It makes coefficient estimates imprecise and unreliable inflates standard errors and diminishes t-statistics making it difficult to distinguish the individual impacts of independent variables.
What is a classic symptom of multicollinearity?
A high R-squared and a significant overall F-statistic but t-statistics for individual slope coefficients that are not significant.
What does the Variance Inflation Factor (VIF) measure?
It quantifies the extent to which an independent variable’s variance is inflated due to its correlation with other independent variables in the model.
What is a common rule of thumb for interpreting VIF values?
A VIF greater than 5 warrants further investigation and a VIF greater than 10 indicates serious multicollinearity requiring correction.
What are possible solutions for correcting multicollinearity?
Excluding one or more correlated variables using a different proxy for one of the variables or increasing the sample size.
What is omitted variable bias?
It’s a bias in coefficient estimates that arises when an important independent variable is omitted from a regression model.
What happens if an omitted variable is correlated with an included independent variable?
The estimated coefficient of the included variable will be biased and inconsistent and its standard error will also be inconsistent.
What is a high-leverage point in regression analysis?
A data point having an extreme value for an independent variable.
What is an outlier in regression analysis?
A data point having an extreme value for the dependent variable relative to its predicted value resulting in a large residual.
What does the leverage measure (hii) indicate and what is a rule of thumb for its use?
Leverage measures the distance of an observation’s independent variable value from the mean. If hii > 3(k+1)/n it is a potentially influential high-leverage point.
What are studentized residuals used for and what is a rule of thumb?
To identify influential outlying Y observations. If the absolute value of the studentized residual exceeds the critical t-value for n-k-2 degrees of freedom the observation is potentially influential.