Final Exam Flashcards

(82 cards)

1
Q

The classical assumptions must be

A

met in order for OLS estimators to be the best available

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Classical Assumption #1

A

The regression model is linear, is correctly specified, and has an additive error term

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Classical Assumption #2

A

The error term has a zero population mean

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Classical Assumption #3

A

All explanatory variables are uncorrelated with the error term

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Classical Assumption #4

A

Observations of the error term are uncorrelated with each other (no serial correlation)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Classical Assumption #5

A

The error term has a constant variance (no heteroskedasticity)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Classical Assumption #6

A

No explanatory variable is a perfect linear function of any other explanatory variable(s) (no perfect multicollinearity)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Classical Assumption #7

A

The error term is normally distributed

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Omitted Variable Bias (Conditions)

A

relevant (β2 ≠ 0) – X1 and X2 are correlated

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Expected bias

A

Expected bias in መ 𝛽1 has two components: the sign of β2 and the sign of Corr(X1, X2)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Limited Dependent Variables

A

We have discussed dummy variables (indicator variables, binary variables) as a tool for measuring qualitative/categorical independent variables (gender, race, etc.)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

linear probability model

A

simply running OLS for a regression, where the dependent variable is a dummy (i.e. binary) variable:
where Di is a dummy variable, and the Xs, βs, and ε are typical independent variables, regression coefficients, and an error term, respectively

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

the term, linear probability model

A

comes from the fact that the right side of the equation is linear while the expected value of the left side measures the probability that Di = 1

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Some issues with LPM

A

෠ 𝑌 𝑖 ≤ 0 or ෠ 𝑌 𝑖 ≥ 1 a more fundamental problem with the linear probability model: nothing in the model requires ෠ 𝑌 to be between 0 and 1!  If ෠ 𝑌 is not between 0 and 1, how do we interpret it as a probability?  A related limitation is that the marginal effect of a 1-unit increase in any X is forced to be constant, which cannot possibly be true for all values of X  E.g., if increasing X by 1 always increases Y by a particular amount, ෠ 𝑌 must exceed 1 when X is sufficiently large

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

The Binomial Logit Model

A

The binomial logit is an estimation technique for equations with dummy dependent variables that avoids the unboundedness problem of the linear probability model

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Logits cannot be estimated using OLS

A

but are instead estimated by maximum likelihood (ML), an iterative estimation technique that is especially useful for equations that are nonlinear in the coefficients  Again, for the logit model is bounded by 1 and 0

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

Basic Procedure for Random Assignment Experiments

A

 Recruit sample of subjects  Randomly assign some to treatment group and some to control group ◼ Random assignment makes treatment uncorrelated with individual characteristics  Measure average difference in outcomes between treatment and control groups

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

natural experiments (or quasi-experiments

A

attempt to utilize the “treatment-control” framework in the absence of actual random assignment to treatment and control groups

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

Difference-in-difference estimator:

A

Policy impact = (Tpost – Tpre) – (Cpost– Cpre)  T: treatment group outcome, C: control group outcome  The DD estimate is the amount by which the change for the treatment group exceeded the change for the control grou

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

Panel data:

A

repeated observations of multiple units over time (combination of cross-sectional and time-series)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

Main advantages of panel data

A

 Increased sample size  Ability to answer types of questions that cross-sectional and time-series data cannot accommodate  Enables use of additional methods to eliminate omitted variables bias

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

Panel Data Notation

A

 𝑖 subscript indexes the cross-sectional unit (individual, county, state, etc.)  t subscript indexes the time period in which the unit is observe

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

First-differenced estimator

A

∆𝑌 𝑖 = 𝛼0 + 𝛽1∆𝑋𝑖 + ∆𝜀𝑖

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
24
Q

Advantages of random effects estimator (if assumption about 𝑎𝑖 is correct):

A

 Allows time-invariant regressors to be included  More degrees of freedom (only estimates parameters of distribution from which 𝑎𝑖 is assumed to be drawn; fixed effects estimator uses one degree of freedom per fixed effect)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
25
Disadvantages of random effects estimator:
 Biased if assumption that 𝑎𝑖 is uncorrelated with regressors is incorrect (while FE estimator allows arbitrary correlation between 𝑎𝑖 and regressors)
26
❑ Fixed effects estimator widely preferred when regressors of interest are time-varying
 It rarely seems likely that 𝑎𝑖 is uncorrelated with any regressors; fixed effects model is generally far more convincing
27
Hausman test
❑ Fixed and random effect estimators can be compared with a Hausman test (previously seen in instrumental variables context as test for endogeneity)
28
Fixed vs. random effects - General Advice
 Under random effects hypothesis, both RE and FE estimators are consistent (should give similar results); under alternative hypothesis, FE consistent but RE is not  Therefore, if estimates are significantly different, can reject null hypothesis of random effects
29
Fixed vs. random effects - Concept
❑ General advice: use fixed effects estimator if it’s feasible
30
T-test
Divide the coefficient by the standard error to get the t-value
31
Omitted Variables – Bias Assessment
Sign (β2) * Sign (Corr (X1, X2)) = Sign of Bias (β1)
32
Irrelevant Variables - Inclusion Criteria
Theory: is there sound justification for including the variable? Bias: do the coefficients for other variables change noticeably when the variable is included? T-Test: is the variable’s estimated coefficient statistically significant? R-square: has the R-square (adjusted R-square) improved?
33
Serial Correlation
First-order serial correlation occurs when the value of the error term in one period is a function of its value in the previous period; the current error term is correlated with the previous error term.
34
DW Test
compare DW(d) to the critical values (𝐝_𝐋, 𝐝_𝐔)
35
Pure Heteroskedasticity
occurs in correctly specified equations
36
Impure Heteroskedasticity
arises due to model misspecification
37
Multicollinearity
Multicollinearity exists in every equation & the severity can change from sample to sample.  There are no generally accepted true statistical tests for multicollinearity. VIF > 5 as a rule of thumb
38
Binary Dependent Variable Models
Linear Probability Model (LPM) & Logit / Probit Model
39
Linear Probability Model (LPM)
- Similar to OLS regression - R-squared is no longer an accurate goodness-of-fit measure - Interpretation: probability that Y=1 on a percentage point scale
40
Logit / Probit Model
- Restricted between 0 and 1 - Automatically corrects for heteroskedasticity - Marginal effect of X is not constant - Not linear in the coefficients
41
LPM Interpretation (example)
On average, a 1-unit increase in DISTANCE is associated with a negative 7.2 percentage point change in the probability of choosing Cedars Sinai, holding all else constant
42
LPM Limitation #1
Unboundedness | The linear probability model produces nonsensical forecasts (>1 and <0)
43
LPM Limitation #2
Adj-R^2 is no longer accurate measure of overall fit
44
LPM Limitation #3
Marginal Effect (slope) of a 1-unit increase in X is forced to be constant
45
LPM Limitation #4
Error term is neither homoskedastic nor normally distributed
46
Logit Model Interpretation - | Coefficient Interpretation:
The sign and significance can be interpreted just as in linear models. B1 is the effect of a 1-unit increase in X1 on the log-odds ratio
47
Calculating Marginal Effects
margin, dydx (X) at mean
48
LPM Interpretations
B*100 percentage point change in the probability that Y=1
49
Logit Interpretations
B change in the log-odds ratio of Y=1
50
Marginal Effects Interpretations
(dy/dx)*100 percentage point change in the probability that Y=1
51
Experimental Methods - | Selection Problems
Treatment is not necessarily randomly assigned because of other systematic differences in the error term (endogeneity) which would cause bias in the treatment’s effect. We need a valid counterfactual to truly understand the effect of the intervention / treatment.
52
Valid Counterfactual
a control group that is exactly the same as the treatment group except it does not receive the treatment
53
Solution to Selection Problems
Randomization | -Researcher randomly assigns subjects to either a treatment or control group to estimate the treatment effect
54
Natural / Quasi-Experiments
Randomized experiments are hard to do in the social sciences, so researchers often rely upon natural experiments where an exogenous event mimics the treatment and control group framework in the absence of actual random assignment
55
Counterfactual Challenge
Counterfactual Challenge | Hard to find an untreated group that really is otherwise identical to the treated group
56
Panel Data - definitions expanded
Formed when cross-sectional and time-series data sets are combined to create a single data set. Main reason for working with panel data (beyond increasing sample size) is to provide insight into analytical questions that can’t be answered by using time-series or cross-sectional data alone
57
Panel Data Advantages
Increased sample size so more degrees of freedom & sample variability Able to answer new research questions Can eliminate omitted variable bias with fixed effects (controlling for unobserved heterogeneity)
58
Panel Data Concerns
Heteroskedasticity & Serial Correlation
59
Panel Data - Fixed Effects Model
Does a good job of estimating panel data equations, and it also helps avoid omitted variable bias due to unobserved heterogeneity.
60
Fixed Effects Model Assumptions
Each cross sectional unit has its own intercept. A fixed effects analysis will allow arbitrary correlation between all time-varying explanatory variables and 𝑎_𝑖
61
Fixed Effects Model Drawback
measurement error, autocorrelation, heteroskedasticity
62
Fixed Effects Model
The omitted variable bias arising from unobserved heterogeneity can be mitigated with panel data and the fixed effects model.
63
How Fixed Effect Model address Omitted Variable Bias
How? Estimates panel data by including enough dummy variables to allow each cross-sectional of individual i and time period t to have a different intercept. These dummy variables absorb the time-invariant, individual-specific omitted factors in the error term
64
Panel Data - Random Effects Model | When to use
When the explanatory variable of interest is time-invariant
65
Panel Data - Random Effects Model | Assumption
Ai and regressors (Xit) are uncorrelated
66
Panel Data - Random Effects Model | Advantages
- Can handle time-invariant variables | - Uses fewer degrees of freedom than FE because of the lack of subject dummies
67
Hausman Test
Compares fixed and random effect estimators to see if their difference is statistically significant. If different  fixed effects model preferred (reject the null hypothesis of random effects) If not different  random effects model to conserve degrees of freedom (or provide estimates of both the fixed effects and random effects models) If both models predict VERY DIFFERENT results, it suggests RE model has omitted variable bias and endogeneity present, making FE more statistically accurate
68
What is the purpose of determining the Cook’s D, what is it used to detect?
Cook’s D is used to detect influential outliers.
69
Linear Probability Model Example
𝐻𝑆𝑖 = 0.42 + 0.028𝑚𝑒𝑑𝑢𝑐𝑖 + 0.002𝑚𝑒𝑑𝑢𝑐𝑖 | 2 + 0.06𝑤𝑜𝑟𝑘�
70
Problems with Linear Probability Model
1) HS෢i could be ≤ 0 or ≥ 1. Linear probability model is difficult to interpret as a probability because HS෢i is not bounded by 0 and 1. The linear probability model produces nonsensical forecasts (greater than 1 and less than 0). 2) The marginal effect of a 1-unit increase in any X is forced to be constant, which cannot possibly be true for all values of X. 3) 𝑅ത2 is no longer an accurate goodness-of-fit measure. The predicted values of Y are forced to change linearly with X, so you could obtain a low 𝑅ത2 for an accurate model.
71
What is the difference between a random and natural experiment?
Random experiments involve researcher’s randomly assigned subjects to either a treatment or control group to estimate the treatment effect. Natural experiments or quasi-experiments, attempt to utilize the “treatment-control” framework in the absence of actual random assignment to treatment and control groups. Instead of researcher randomly assigning treatment, rely on some exogenous event to create treatment and control groups. When the event or the policy is truly exogenous, treatment is as good as randomly assigned.
72
Problems in random experiments.
1) Random experiments are often very costly or cannot be carried out due to being unethical. 2) Non-random samples. They often lack generalizability since the sample may not be randomly drawn from the entire population of interest. 3) Attrition bias because treatment or control units non-randomly drop out of the experiment. 4) Hawthorne effects (people behave differently when observed, may respond to treatment/control status). 5) Randomization failure (can only control for observed treatment-control differences; bias may result if unobservable characteristics not perfectly balanced).
73
Explain briefly the difference-in-differences estimator
This method estimates the impact of a treatment by comparing the outcomes of a treatment group and a control group before and after the treatment is received.
74
Main underlying assumption of difference-in-differences estimator
The main underlying assumption: in the absence of the treatment, the difference between the outcomes of the two groups would not have changed (i.e., they would have followed a common trend). The change in outcomes of the control group is viewed as the counterfactual for the change in outcomes of the treatment group.
75
What is panel data?
Panel data are repeated observations of multiple units over time. It is a combination of cross-sectional and time-series.
76
Advantages of Panel Data
1) More degrees of freedom and more sample variability than cross sectional data alone or time series data alone allow for more accurate inference of the model parameters and hence increase the efficiency of estimates. 2) Eliminate omitted variables bias. It is often argued that the real reason someone finds an effect is because of ignoring specific variables that are correlated with the explanatory variables when specifying the model. Panel data allows us to control for missing or unobserved variables. 3) Ability to answer types of questions that cross-sectional and time-series data cannot accommodate. For example transitions from employment to unemployment, from employment to retirement, changes on health status or any other variables that can change through time.
77
differences between the fixed effects and random effects panel data models.
The fixed effects model allows for 𝑎𝑖 to be correlated with the regressors and the random effects estimator assumes 𝑎𝑖 is not correlated with the regressors.
78
Advantages of the random effects model
Advantages of the random effects model are that allows time-invariant regressors to be included and it includes more degrees of freedom.
79
Disadvantages of the random effects model
The main disadvantage of random effects estimator is that is biased if the assumption that 𝑎𝑖 is uncorrelated with regressors is incorrect.
80
Advantage of the fixed effects model
An advantage of the fixed effects model is that allows arbitrary correlation between 𝑎𝑖 and any regressors
81
Disadvantage of the fixed effects model
One of the main disadvantages is that drops out time-invariant regressors.
82
Preference between fixed effects model and random effects model
Unless we wish to estimate the effect of a time invariant variable, fixed effects are generally preferred over random effects due to having less restrictive assumptions.