HANDOUT 5 Flashcards

(57 cards)

1
Q

Panel data allows us to…

A

observe multiple individuals over multiple points in time.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

restriction on individuals for panel data

A

MUST be the SAME individuals that we follow over time.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Strongly balanced data =

A

same number of observations for every individual

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Why may the data be unbalanced? Is this an issue?

A

Some individuals drop out the study = attrition rate. Problem if attrition rate = f(X variables) –> self-selection bias = sample no longer representative.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

2 types of unobserved heterogeneity

A
  1. unobserved individual heterogeneity = ai

2. unobserved time heterogeneity = dt

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Omitted relevant variable formula for Eit Ai

A

E(b1) = B1 + B2 COV(Ai, Eit) / Var(Eit)

E(b1)≠B1

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Pooled OLS =

A

same intercept, same slope
Use all NT observations
No dummies = no heterogeneity
Treat same individual at 2 points in time as 2 separate individuals.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Pooled OLS equation

A

Yit = alpha + BXit+ €it

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

4 assumptions for pooled OLS

A
  1. Each individual randomly selected
  2. No individual heterogeneity
  3. No time-specific heterogeneity = parameters constant across time (no structural change)
  4. E(€it I Xit) = 0 - strict exogeneity
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Are estimates for pooled OLS good?

A

They’re consistent for large N / large T / large NT.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

beta interpretation for pooled ols

A

average increase in Y for unit increase in X, averaged across all individuals and all times.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

formula for pooled OLS estimate of b

A

sum i=1,..,N sum t=1,…,T (Xit - X bar)(Yit - Y bar)

/ sum i=1,..,N sum t=1,…,T (Xit - X bar)^2

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

How can we have different intercepts and different slopes for all individuals?

A

Yit = alpha i + Bi Xit + €it

Run N separate regressions - get alpha i and bi for every individual. get sigma^2 for every individual.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

We can only do OLS with different intercepts and different slopes if…

A

T is large enough for every i = consistency.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

bi formula for different slopes, different intercepts

A

bi = sum t=1,..,T (Xit - Xi bar)(Yit - Yi bar) /
sum t=1,..,T (Xit - Xi bar)^2
De-mean using individual’s own time average.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

How else can we run a regression that allows intercepts and slopes to differ for all individuals?

A

Run one large regression with additive and multiplicative dummies for N-1 individuals - this way we only get one sigma^2 estimate.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

Fixed effects / within-groups model allows…

A

Different intercepts, same slopes

= allows for individual heterogeneity ai

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

Equation for FE model

A

Yit = alpha + BXit + ai + €it

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

b formula for FE

A

b = sum i=1,..,N sum t=1,…,T (Xit - Xi bar)(Yit - Yi bar)

/ sum i=1,..,N sum t=1,…,T (Xit - Xi bar)^2

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

Why is FE model also know as “within-groups” model?

A

Because we de-mean by an individuals own time average (then average across all individuals)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

How many parameter estimates do we get for FE?

A

1 x slope coefficient Beta

N x intercept estimates

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

Slope coefficients in FE are consistent if…

A

Large N / large T / large NT

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

Intercepts in FE are consistent if…

24
Q

De-meaning method for FE

A

Yit = alpha + B1Xit + B2Ai + €it
Yi bar = alpha + B1 Xi bar + B2 Ai + €i bar
(Yit - Yi bar) = B1(Xit - Xi bar) + (€it - €i bar)

25
Does it matter if we omit Ai for FE model?
NO - because de-meaning eliminates Ai anyway = we can control for unobserved individual heterogeneity. So don't worry about bias.
26
3 stages of partitioned regression for FE
1. Regress Yit on Ai & save resid (Yit - Yibar) 2. regress Xit on Ai & save resid 3. Regress Yi tilda on Xi tilda = we've stripped out the individual heterogeneity.
27
In a FE model, our X variables must...
MUST vary with time. So we cannot estimate coefficients on race, gender etc. we can only control for them as part of FE.
28
Ai captures...
all the time invariant variables.
29
Random effects model equation
Yit = alpha + BXit + ai + €it
30
How does RE differ to FE? What 3 assumptions do we make?
WE make some assumptions about ai, which are random drawings from a distribution: E(ai) = 0 V(ai) = sigma^a - time invariant COV(ai, aj) = 0
31
In a RE model, what happens to ai?
It becomes part of the error term | Ui = ai + €i
32
Does Ui in RE model satisfy CLRM assumptions?
1. E(Ui) = 0 yes 2. V(Ui) = sigma^2 a + sigma^2 yes time invariant 3. COV(Uit, Uis) = sigma^2 a ≠ 0 violates CLRM.
33
Why is there a non-zero covariance between Uit and Uis for RE?
Looks like serial correlation | Because ai is time invariant and is part of error term.
34
Solution to serial correlation in RE model
Transform the equation so that the error term is serially uncorrelated within i.
35
lamba to transform RE model =
Lambda = 1 - sqrt[sigma^2 / (sigma^2 + Tsigma^2 a)]
36
Transformed RE model equation
(Yit - lamba Yi bar) = (1-lambda) alpha + | B(Xit - lamba Xibar) + (Rit - lamba Ri bar)
37
If lamba=0, what does RE model become?
lamba = 0 --> POOLED OLS If lambda = 0, sigma^2 a = 0 So NO individual heterogeneity Yit = alpha + BXit + Rit
38
If lamba=1, what does RE model become?
lamba = 1 --> FE Model If lambda = 1, sigma^2 a --> infinity HUGE individual heterogeneity Yit - Yi bar = B(Xit - Xi bar) + (Rit - Ri bar)
39
If 0 < lambda < 1, which model is most efficient?
RE model
40
FE vs RE when COV(ai, Xit) = 0
``` RE = unbiased, efficient FE = unbiased, but inefficient ```
41
FE vs RE when COV(ai, Xit) ≠ 0
``` RE = efficient still, but BIASED FE = unbiased still, but inefficient ```
42
Why is FE model inefficient? But then why is it useful?
inefficient as estimate a lot of parameters - different intercept for every individual. But good if huge individual heterogeneity.
43
Test for FE vs RE
Hausman Test H0: COV(ai, Xit) = 0 - RE correct and efficient H1: COV(ai, Xit)≠0 - RE incorrect FE correct under either
44
Test stat for Hausman test
H = (b FE - b RE)^2 / V(b FE) - V(b RE)
45
Hausman test: CVs from...
Chi-squared distribution | dof = number of slope coefficients we estimate
46
Hausman test stat under H0 and H1
H0: b FE = b RE so H --> 0 H1: b FE ≠ b RE so H --> infinity as we square Under both V(b FE) > V(b RE) so denom > 0
47
We can also do the Hausman test for...
OLS vs IV H0: COV(Xi, €i) = 0 OLS is efficient, but biased under H1 IV is always ok, but inefficient
48
First difference model equation
change Yt = B change Xt + change €it | Ai eliminated.
49
In first difference model, t goes from...
t = 2,...,T | 2 because we have a first difference so decrease no observations by 1.
50
OLS estimate of b for first difference model
b = sum i =1,...,N t=2,...,T (changeXit - changeX bar)(changeYit - change Y bar) / sums (change Xit - change X bar)^2
51
In first difference model, change X bar =
change X bar = sums change Xit / n(T - 1)
52
When is first difference model same as FE?
when T=2
53
When is first difference model NOT same as FE?
When T>2
54
When T>2, worry with first difference model =
The error term is MA(1) process | change €it = €it - €it-1
55
Complications with dynamic model for panel data
OLS -> biased estimators As t-->infinity, and N fixed, estimates consistent But usually in panel data T small & large N
56
What is the bias caused by in dynamic model?
BY having to eliminate ai from each observation --> correlation of order (1/T) between lagged dependent variable and residuals.
57
Solution to OLS bias in dynamic model
Take first differences & then use instruments to do IV estimation. Since error term = MA1 = 1 period memory, use lags yt-2 or anything before as replacement for yt-1.