QM 2 Flashcards

(61 cards)

1
Q

What is conditional heteroskedasticity?

A

It occurs when the error variance in a regression is correlated with or conditional on the values of the independent variables.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What are the main consequences of conditional heteroskedasticity for statistical inference?

A

Standard errors of regression coefficients are underestimated t-statistics are inflated and Type I errors rejecting a true null hypothesis are more likely. The F-test for overall significance also becomes unreliable.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is the purpose of the Breusch-Pagan (BP) test?

A

To test for conditional heteroskedasticity in a regression model.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is the null hypothesis of the Breusch-Pagan (BP) test?

A

There is no conditional heteroskedasticity meaning the regression’s squared residuals are uncorrelated with the independent variables.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

How is the Breusch-Pagan (BP) test statistic calculated and what is its distribution?

A

The BP test statistic is n multiplied by R-squared from a regression of the squared residuals on the original independent variables. It is approximately chi-square distributed with k degrees of freedom where k is the number of independent variables.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

How do you correct for conditional heteroskedasticity?

A

By computing robust standard errors also known as heteroskedasticity-consistent or White-corrected standard errors.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is serial correlation or autocorrelation in a regression model?

A

It occurs when regression errors are correlated across observations typically in time-series regressions.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What are the consequences of serial correlation if one of the independent variables is a lagged value of the dependent variable?

A

All parameter estimates (coefficients) become inconsistent and invalid. Standard errors are also invalid.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What are the consequences of positive serial correlation if no independent variable is a lagged dependent variable?

A

Coefficient estimates are consistent but standard errors are typically underestimated leading to inflated t-statistics and more Type I errors. The F-statistic may also be inflated.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What is the Breusch-Godfrey (BG) test used for?

A

To detect serial correlation of a pre-designated order p in regression residuals.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is the null hypothesis of the Breusch-Godfrey (BG) test?

A

There is no serial correlation in the model’s residuals up to the specified lag p.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

How are robust standard errors (like Newey-West) useful in the context of serial correlation?

A

They adjust the coefficient standard errors to account for serial correlation and also correct for conditional heteroskedasticity allowing for more reliable statistical inference.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What is multicollinearity in a multiple regression model?

A

It occurs when two or more independent variables are highly correlated or when there is an approximate linear relationship among independent variables.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What are the primary consequences of multicollinearity?

A

It makes coefficient estimates imprecise and unreliable inflates standard errors and diminishes t-statistics making it difficult to distinguish the individual impacts of independent variables.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What is a classic symptom of multicollinearity?

A

A high R-squared and a significant overall F-statistic but t-statistics for individual slope coefficients that are not significant.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What does the Variance Inflation Factor (VIF) measure?

A

It quantifies the extent to which an independent variable’s variance is inflated due to its correlation with other independent variables in the model.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

What is a common rule of thumb for interpreting VIF values?

A

A VIF greater than 5 warrants further investigation and a VIF greater than 10 indicates serious multicollinearity requiring correction.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

What are possible solutions for correcting multicollinearity?

A

Excluding one or more correlated variables using a different proxy for one of the variables or increasing the sample size.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

What is omitted variable bias?

A

It’s a bias in coefficient estimates that arises when an important independent variable is omitted from a regression model.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

What happens if an omitted variable is correlated with an included independent variable?

A

The estimated coefficient of the included variable will be biased and inconsistent and its standard error will also be inconsistent.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

What is a high-leverage point in regression analysis?

A

A data point having an extreme value for an independent variable.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

What is an outlier in regression analysis?

A

A data point having an extreme value for the dependent variable relative to its predicted value resulting in a large residual.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

What does the leverage measure (hii) indicate and what is a rule of thumb for its use?

A

Leverage measures the distance of an observation’s independent variable value from the mean. If hii > 3(k+1)/n it is a potentially influential high-leverage point.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
24
Q

What are studentized residuals used for and what is a rule of thumb?

A

To identify influential outlying Y observations. If the absolute value of the studentized residual exceeds the critical t-value for n-k-2 degrees of freedom the observation is potentially influential.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
25
If you want to distinguish among n categories using dummy variables how many dummy variables should you include in the regression?
You should include n-1 dummy variables. The nth category becomes the base or control group.
26
What does an intercept dummy variable do in a regression model?
It adds to or reduces the original intercept if a specific condition (represented by the dummy being 1) is met shifting the regression line up or down.
27
What does a slope dummy variable do in a regression model?
It allows for a change in the slope of the relationship between an independent variable and the dependent variable if a specific condition (represented by the dummy being 1) is met.
28
When is logistic regression typically used?
When the dependent variable is qualitative or categorical often binary (e.g. event happens or does not happen).
29
What transformation is applied to the dependent variable in logistic regression?
The logistic transformation which converts the probability P of an event into the log-odds ln(P/(1-P)).
30
How are slope coefficients in a logistic regression model interpreted?
As the change in the log-odds that the event happens per unit change in the independent variable holding all other independent variables constant.
31
What is the Likelihood Ratio (LR) test used for in logistic regression?
To assess the fit of logistic regression models by comparing the log-likelihood of a restricted model to an unrestricted model similar to an F-test in OLS regression.
32
What does a pseudo-R-squared in logistic regression measure?
It attempts to capture the explained variation similar to R-squared in OLS but must be interpreted with care and is mainly for comparing different specifications of the same model.
33
What is overfitting in a machine learning model?
When a model fits the training data too well capturing noise and random fluctuations such that it does not generalize well to new out-of-sample data.
34
What is the difference between bias error and variance error in ML models?
Bias error is the degree to which a model fits training data (high bias means underfitting). Variance error is how much model results change with new data (high variance means overfitting).
35
What is k-fold cross-validation used for?
To estimate a model's out-of-sample error and mitigate overfitting by repeatedly splitting the data into training and validation sets.
36
What is the primary goal of penalized regression techniques like LASSO?
To prevent overfitting by adding a penalty term to the sum of squared residuals that increases with the number of included features or the magnitude of coefficients.
37
What does LASSO (Least Absolute Shrinkage and Selection Operator) do?
It minimizes the sum of squared residuals plus a penalty term based on the sum of the absolute values of coefficients effectively shrinking some coefficients to zero and performing feature selection.
38
What is a Support Vector Machine (SVM) primarily used for in supervised learning?
For classification tasks aiming to find the optimal hyperplane that separates data points into distinct classes with the maximum margin.
39
How does the K-Nearest Neighbor (KNN) algorithm classify a new observation?
By finding the k most similar (nearest) observations in the existing labeled dataset and assigning the new observation to the majority class among those k neighbors.
40
What are Classification and Regression Trees (CART) used for?
For both classification (predicting a categorical target) and regression (predicting a continuous target) by partitioning data into subgroups based on feature values.
41
What is ensemble learning in machine learning?
A technique of combining the predictions from a collection of models to typically produce more accurate and stable predictions than any single model.
42
What is a random forest classifier?
An ensemble learning method that constructs a multitude of decision trees during training and outputs the class that is the mode of the classes (classification) or mean prediction (regression) of the individual trees.
43
What is Principal Components Analysis (PCA) used for in unsupervised learning?
For dimension reduction by transforming a set of correlated features into a smaller set of uncorrelated composite variables (principal components) that retain most of the original variance.
44
What is the goal of clustering algorithms like K-Means?
To sort observations into groups (clusters) such that observations within the same cluster are more similar to each other than to observations in other clusters.
45
How does K-Means clustering work?
It partitions observations into a pre-specified number (k) of clusters by iteratively assigning observations to the nearest cluster centroid and then recalculating the centroids.
46
What is hierarchical clustering?
An unsupervised algorithm that builds a hierarchy of clusters either agglomeratively (bottom-up starting with individual observations) or divisively (top-down starting with one large cluster).
47
What is a neural network (NN) in machine learning?
A flexible algorithm inspired by the human brain with layers of interconnected nodes (neurons) that transform inputs non-linearly to learn complex patterns.
48
What distinguishes a Deep Neural Network (DNN) from a shallow neural network?
A DNN has multiple hidden layers (typically at least 2) allowing it to learn more complex hierarchical features from data.
49
What is reinforcement learning (RL)?
A type of machine learning where an agent learns to make decisions by performing actions in an environment to maximize a cumulative reward often through trial and error.
50
What is the primary challenge of working with time-series data in regression?
The assumptions of linear regression such as uncorrelated errors and constant variance (homoskedasticity) are often violated.
51
What is a covariance stationary time series?
A time series whose mean variance and autocovariances are constant and finite over time.
52
What happens if you estimate an autoregressive (AR) model for a time series that is not covariance stationary?
The estimation results (coefficients standard errors) will likely be invalid and have no economic meaning.
53
How can you test if the residuals of an AR model are serially correlated?
By examining the t-statistics of the residual autocorrelations. The Durbin-Watson test is generally invalid for AR models.
54
What is mean reversion in a time series?
A tendency for the series to fall when its level is above its long-run mean and rise when its level is below its mean. Covariance stationary series are mean reverting.
55
What is the root mean squared error (RMSE) used for in time-series forecasting?
To compare the out-of-sample forecasting accuracy of different models. A smaller RMSE indicates greater accuracy.
56
What is a random walk process?
A time series where the value in one period is the value in the previous period plus an unpredictable random error. It is not covariance stationary.
57
What is a unit root in a time series?
A characteristic of a non-stationary time series (like a random walk) where the coefficient on the first lag in an AR(1) model is equal to 1.
58
How can a time series with a unit root often be transformed to achieve stationarity?
By first-differencing the time series (i.e. taking the difference between consecutive observations).
59
How can seasonality in a time-series model often be addressed?
By including seasonal lags in an autoregressive model (e.g. adding a 4th lag for quarterly data or a 12th lag for monthly data).
60
What does Autoregressive Conditional Heteroskedasticity (ARCH) model?
It models the variance of the error term in a time series as dependent on the variance of previous errors meaning volatility can cluster.
61
If two time series both have unit roots when can linear regression between them be valid?
If the time series are cointegrated meaning there is a long-term stable economic relationship between them and their linear combination is stationary.