Final Flashcards

(58 cards)

1
Q

What is Ŷ?

A

An estimated value of Y calculated from the regression at the i-th observation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What can cause stochastic error?

A
  1. OVB 2. Measurement error 3. A misspecified function 4. Random occurrences
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What can we use to quantify the severity of multicollinearity in our model?

A

We use the variance inflation factor (VIF)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is the VIF(ß1 hat?)

A

1/(1-R2)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Which hypothesis test do we use for the following situations?: 1. Single parameter (one restriction) 2. Multiple parameter (linear combination) 3. Multiple parameter (non-linear combination) 4. Multiple parameter (multiple restrictions)

A
  1. t-test (of one variable) 2. lincom (or t-test comparing two variables) 3. t-test/Taylor approximation (if one restriction) 4. F-test
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What does ß0 equal? (Univariate)

A

Ybar - ß1(Xbar)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

How do you know if you have OVB?

A
  1. Use economic theory/knowledge of the subject 2. Run a robustness check
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is TSS?

A

Sum of the squared differences between actual values and the mean

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What is RMSE?

A
  1. Another goodness of fit measurement 2. Sqrt of SSR/(n-k-1)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What is the advantage of BIC over AIC?

A

BIC gives you consistent estimates

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What should you do if you are asked about an effect? A change?

A

Effect = derivative

Change = difference

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What is a residual?

A

The difference between the estimated and actual values of the dependent variable

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What is R2?

A

The % of variation in Y explained by the model

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What are the four assumptions of the error term?

A
  1. The variation does not change as X changes 2. Its distribution is normal 3. Conditional mean=0 4. Independent for any two observations (i,j)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What is the formula for the adj. R2?

A
  1. 1 - [(n-1)/(n-k-1)] x SSR/TSS
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What do you need to look at after running an F-test to determine if it is statistically significant?

A

The Chi-square critical values

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

What does ß1 equal? (Univariate)

A

∑(Yi-Ybar)(Xi-Xbar)/∑(Xi-Xbar)2

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

What is the formula for an F-test?

A

(SSRr - SSRu/r) ÷ (SSRu/n-k-1)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

If our causal effect depends on the level of another independent variable, what do we do?

A

Take the natural log of the variable

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

What must “t” be greater than equal to for: 1. 90% confidence 2. 95% confidence 3. 99% confidence

A

1) 1.645 2) 1.96 3) 2.58 ^^ For two-tailed tests

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

If our causal effect depends on another variable (but not the level) what do we do?

A

Use an interaction terms

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

Formula for t-test?

A

1 hat)-(H0: ß1) / SE(ß1 hat) p = 2(cdf)(-l t l)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

What is the stochastic error term?

A

A term added to the regression equation to account for any variation in Y that is not explained by X

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
24
Q

What are the three properties of estimators?

A
  1. Unbiasedness: The estimator is correct (on avg.) 2. Consistency: As observations increase, so does the probability that the estimator is close to the pop. parameter Efficiency: Estimator has smaller relative variance (converges to the pop. parameter more quickly)
25
For a hypothesis test, what is the significance level and confidence level?
Significance = p Confidence = 1-p
26
What is iteratively re-weighted least squares?
Feasible GLS repeated until weights converge to a value
27
While adding a variable may not change TSS...
It will likely reduce SSR and, thus likely increase R-squared
28
OLS seeks to minimize....
The sum of squared residuals (or SSE)
29
Why is a high degree of freedom desired?
It is likely that the errors will balance out
30
What type of GLS do we use when the form of heteroskedasticity is unknown?
Feasible GLS
31
If the omitted variable is correlated with a regressor and it has an effect on the dependent variable....
We have OVB
32
When are dummy variables useful?
When we want to quantify something that is inherently qualitative (race, gender, etc.)
33
What assumption do we make about the causal effect if we use a natural log?
That it is always positive or negative
34
What property about our OLS estimator is violated if we have heteroskedasticity?
Efficiency
35
How do we adjust for degrees of freedom?
Divide by (n-1)
36
If the effect of the OV on Y and the correlation between OV and regressor are moving in the same direction...
Your estimator is too big
37
K = ?
The # of independent variables
38
What is the SE?
Sq.Rt. of RSS/{(n-k-1) ∑(Xi-Xbar)2}
39
If you specify a dummy variable for each possible outcome.....
You will induce perfect multicollinearity (nothing to compare dummy to)
40
What does TSS equal?
ESS+SSR
41
What are the formulas for R2?
1. ESS/TSS 2. 1 - SSR/TSS
42
What happens if we have perfect multicollinearity?
OLS is impossible
43
What changes between observations?
The values of Y, Xs, and error terms (but not the coefficients)
44
What is the formula for sample variance?
1/(n-1) ∑(Xi-Xbar)2
45
What is ESS?
Sum of the squared differences between predicted values and the mean
46
How do we run feasible GLS?
1. Estimate w/ OLS and calculate residuals 2. Run OLS w/ squared residuals on variance 3. Use predicted values from that to create weight (1/sq.rt(predicted values))
47
What is the formula for sample covariance?
1/(n-1) ∑(Xi-Xbar)(Yi-Ybar)
48
What do you do if asked about the effect of *x* on *y* for a person w/ *z* years of *x*?
Plug *z* into *x* and solve \*\*(Final answer \* Ɛ)
49
What is the extensive formula for ESS?
1)2 ∑(Xi-Xbar)2
50
Why is perfect (or high) multicollinearity an issue?
A high degree of multicolinearity may be problematic because it inflates the variance of the estimator
51
What do you do if asked which level of *x* has a max effect on *y*?
Take the derivative and solve for *x*
52
What do you do if asked about the difference in *y* due to a difference in *x*?
Plug in given values and take the difference
53
How do you standardize a normal distribution?
Subtract the mean and divide by sigma
54
What are the 7 OLS assumptions?
1. The population regression function (DGP) is linear in parameters 2. Observations are randomly drawn from the population and i.i.d 3. X[vector] is fixed in repeated samples (no measurement error) 4. The error term has a conditional mean of 0 5. Homoskedasticity 6. Errors are independent (for every i, j) 7. Outliers are unlikely
55
Yi - Ŷ is....
The residual (prediction mistake)
56
What is the extensive formula for R2?
1) ∑(Xi-Xbar)(Yi-Ybar)/∑(Yi-Ybar)2
57
What do you do if asked about the difference in the effect of *x* on *y*? (Someone w/ 10 years vs. someone w/ 20)
Take the derivative, plug in the numbers and take the difference
58
What are the advantages of using a non-linear specification other than a log?
No assumptions about direction, allows for inflection points, and increasing/decreasing rates