Binary Choice/Censored Data Flashcards

(12 cards)

1
Q

What major flaw does the linear probability model have?
E(yi|xi)=P(yi=1|xi)=xi’β;
Mean E(yi|xi)=P(yi=1|xi)=xi’β
Variance: Var(yi|xi)=P(|xi)(1-P(|xi))=xi’β(1-xi’β)

A

It assumes the conditional probability function to be linear. This does not restrict P(y=1|x1,…,xk) to lie between 0 and 1. Thus you can get probabilities lower than 0 and higher than 1. Therefore we need probit/logit regression

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

E(yi|xi)=F(xi’β0): then probit:

A

F is the standard normal distribution: F(t)=φ(t)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

E(yi|xi)=F(xi’β0): then logit:

A

F is the standard logistic distribution with the location parameter 0 and scale parameter 1: F(t)=exp(t)/(1+exp(t))=1/(1+exp(-t))

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Coefficients in the probit/logit models cannot be interpreted in the same way as in the linear regression models, because:

A

probit/logit only reveal the signs of the effect, not the magnitude. However, when we look at marginal effects, interpretation is possible. This marginal effect is reported either at the sample mean of xi or the average marginal effect is taken

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Marginal effect (continuous variables):

A

∂E(yi|xi1,…,xip)/∂xij=∂P(yi=1|xi1,…,xip)/∂xij

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Marginal effect (discrete variables):

A

P(yi=1|xi1,…,xij+1,…,xip)-P(yi=0|xi1,…,xij,…,xip)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Marginal effect of variable xij:

A

pj(x)=∂P(yi=1|xi1,…,xip)/∂xij=∂F(x’β)/∂xij=f(x’β)βj

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Measures of fit:

A

1: Percentage correctly predicted (PCP)=ΣI(yi=yi^)/n where yi^=I(F(x’βn^)>0.5)
2: Pseudo-R^2=1-ln(Ln(βn^)/ln(Ln((1,0,…,0)’)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Censored data:

A

Some values are not observable. Only the lower bound is known. Transformation to censoring from below at 0 (yi* is the latent/uncensored variable, yi is observable):
yi=max(a,yi) <–> yi-a=max(0,yi-a)
yi=min(a,yi) <–> -yi=max(-a,-yi)
Assume yi=max(0,yi*)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Tobit model

A

In this model, the data is censored. We estimate the model with maximum likelihood:
1: (xi,yi) random sample
2: yi=xi’β+εi; yi=max(0,yi)
3: εi
|xi~N(0,σ^2)
4: (xi,εi*) with E(xixi’)>0 has full rank
5: εi=yi-xi’β

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Tobit model likelihood contribution for yi=0:

A

P(yi=0|xi)=1-φ(xi’β/σ)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Tobit model likelihood contribution for yi>0:

A

f(yi|xi)=(1/σ)φ((yi-xi’β)/σ)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly