L3 Flashcards

1
Q

What is omitted variable bias?

A

When a variable that affects the independent variable is not included in the model specification, and therefore resides in the error term. This causes the error term to become correlated with one/more of the x variables, causing bias (eg. ‘unobserved ability’ in a wage estimation model?)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

2 conditions which the omitted variable ‘Z’ must satisfy to create OVB?

A

1) Z is a determinant of Y (ie. a part of u)

2) Z is correlated with the regressor X (Corr(Z,X) nto equal to 0)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

see

A

slides 5, 6, and 7 for example of OVB

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Which direction will an omitted variable bias go when there is a positive correlation between X and u? and for negative?

A

Upwards

If negative correlation, bias will go in a downward direction

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Explain what an ideal randomised controlled experiment is by defining each term?

A
Ideal = subjects follow the treatment protocol-perfect compliance (ie. no reporting errors etc.)
Randomised = subjects from pop. of interest are randomly assigned to a treatment/control group
Controlled = having a control group allows for measuring differential effect of treatment
Experiment = treatment is assigned as part of an experiment, tf subjects have no 'choice' tf no reverse causality
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Define a causal effect?

A

The effect measured in an ideal randomised controlled experiment

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Will there ever be OVB in an IRCE? Why?

A

Bc E(ui|X)=0 (ie. LSA1 holds) tf no OVB

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

3 ways to overcome OVB?

A

1) Run a randomised experiment where the treatment is assigned - this works because although part of the error still determines Y, it is now uncorrelated with the regressor X (rarely feasible)
2) Use ‘cross-tabulation’ approach
3) Use a regression in which the OV is no longer omitted! (see notes for examples)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Explain what the ‘cross-tabulation’ approach is, and how it controls for OVB? What is a problem with it?

A

Within each group, control for the OVB by having each subgroup have the same level of the OVB (see notes) (may run out of data!)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

MRM: What is beta(i)?

A

The effect on Y of a change in beta(i) holding all beta(j) constant (where i not equal to j)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is beta(0) in the MRM?

A

Predicted value of Y when all Xi=0

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

MRM: What is R^2 and what is R(bar)^2? What is the difference between them?

A
R^2 = fraction of variability in Y explained by X
R(bar)^2 = adjusted R^2; R^2 with a DofF correction that adjusts for estimation uncertainty

Normally, R(bar)^2 will be smaller, but if n is very large they will be roughly the same

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Explain why the adjusted R^2 is more appropriate for a MRM?

A

The normal R^2 will always increase when you all in another regressor; this is a problem for measuring ‘fit’ because almost all data will have some kind of correlation with Y, and the R^2 normal sees this as increased explanatory power of the model.
The adjusted R(bar)^2 corrects this problem by ‘penalising’ you for adding another regressor tf it won’t necessarily increase as you add more Xs

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

State the 4 LSAs for Multiple regression?

A

See notes; side 2 at the bottom

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

MRM: what happens if the LSA 1 fails?

A

This implies there is an OVB (tf either include OV in regression or a variable that controls for the OVB)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

How should one check if LSA 3 holds?

A

Check scatterplots for outliers/typos

17
Q

Define perfect multicollinearity?

A

When one of the regressors is an exact linear function of one of the other regressors

18
Q

What happens if you include k binary regressors for k discrete options?

A

Perfect multicollinearity - should use k-1 regressors

19
Q

What is the Dummy Variable Trap?

A

Given a set of multiple dummy variables that are mutually exclusive and exhaustive, if you include ALL these dummys and a constant beta(0), there will be perfect multicollinearity

20
Q

2 solutions to the DVT?

A

1) Omit one of the groups

2) Omit the constant (intercept)

21
Q

What is imperfect multicollinearity?

A

When 2/more of the regressors are very highly correlated

22
Q

What is the implication of imperfect multicollinearity? Explain why.

A

One or more of the regressors will be imprecisely estimated

Why? β(1) is the effect of X1 on Y holding X2 constant. BUT if X1 and X2 are highly correlated, then there is very little variation in X1 ONCE X2 is held constant tf the data lacks information about the change in X1 when X2 is held constant which leads to large variance of OLS estimator on X1 and large SEs