Chapter 5 - Brooks Flashcards

(31 cards)

1
Q

what is the purpose of diagnostic tests?

A

To see whether the CLRM holds as a valid tool

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

what is the foundation of this chapter+

A

The two types of ways to derive all the tests we are working with.

1) LM approach
2) Wald approach

There is a third approahc, LR /likelihood ratio).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Distinguish LM and Wald test

A

LM has chi-squared variable.

Wald has F-distributed statistic.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

elaborate on LM test

A

follow chi squared variable with degrees of freedom equal to the number of restrctions placed on the restricted regression.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

elaborate on the Wald test

A

F-dist, with F(m, T-k) degrees of freedom

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

what is the relationship between LM and Wald

A

Asymptotically, they are the same. This is because when the sample size grows, the “T-k” degrees of freedom grow towards infinity.

This is becasue of the behavior of the chi squared variable. By definition, it is a sum of standard normal distributed variables (squared). Because they are squared, the expectation grows with the size of the lenght of the sum. The larger the sum, meaning the more degrees of freedom there are, the larger the mean value will be. Recall that the mean is k while variance is 2k.

Anyways, more degrees of freedom implies that the variable value grows. When we consider the F-distribution with denominator that has degrees of freedom that grow with the sample size, this implies that as the sample size grows, the denominator decrease.

Outcome is that asymptocialyl, the F-dist is a chi squared variable.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

elaborate on the first assumption

A

E[u_t] = 0

Expected value of residuals is 0.

The only thing that guarantee this is the inclusion of the intercept term. If we do not have this included, we cannot guarantee that the average error is 0.

In fact, we could get negative R^2 values if we do not have a variable intercept.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

elaborate on the second assumption

A

The second assumption is about the variance of the residuals being constant.

This is commonly referred to as the assumption of homoskedasticity.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

what do we say if residuals are not constant in variance

A

The errors are heteroskedastic

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

how do we detect heteroskedaticiry

A

we consider 2 primary tests.

1) Goldfeld Quandt
2) White’s test

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

elaborate on GQ

A

We split the sample into two subsamples.

The regression model is estimated on each subsample, and then the residual VARIANCES is found for both.

We do not need the regression on the full sample.

The null hypothesis is that variance is constant, so we have sigma^2_1 = sigma^2_2.

The test statistic is the ratio between those two values. This is therefore F(T1-k, T2-k).

Very simple test to conduct, but it is sensitive to the point of break.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

how can goldfeld quandt be improved?

A

increase the size of the mid-region dropout. This will increase the power of the test.

Recall that the power of the test is the probability of making a type 2 error (reject the null hypothesis when it is actually true). Therefore, making the mid-region dropout larger will make it less likely that we say that some value is insignficiant when it is actually true.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

alternative of GQ test

A

White’s test

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

elaborate on white’s test

A

it has the advantage over GQ in the sense that it makes no assumption about the SHAPE of the heteroskedasticity.

This test relies in trying to figure out if there exists a systematic variation in the squared residuals or not.

We make a regular regression, obtain the residuals, and then we square the residuals and fit a NEW regression that has both linear regular coefficients, but also squares the variables etc. We include cross products of the variables as well.

This is based on how variance of residuals reduce to E[u_t^2] because the E[u_t]^2 part is 0 under the validity of the first assumption.

Now we have 2 approaches for the remaining part.

We can basically choose a framework, chi squared or F-dist.

If we use F-dist, we need to find 2 regressions.
We would use the auxillary regression for both, meaning that we try to regress on the squared residuals of the OG regression for both. however, the unrestricted one is the one that has all thhe regular terms, squared terms, cross terms. The restricted one has only a constant. Then we use the hyp that all params are 0, which basically mean that there is no systematic variation that explains the variance.

The LM approach is based on R^2 etc.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

what happens if we use CLRM on a model where assumption 2 is violated?

A

1) The estimator will remain unbiased. Think of what for instance increasing variance looks like on a plot.

2) The estimator is no longer “best” linear. They no longer have the best efficiency.

3) the standard errors can be wrong. This is becasue the standard errors depend on the variance of the data. With wrong standard errors, we dont really know what we get.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

how can we deal with heteroskedasticity?

A

One could try to transform into logs.

17
Q

elaborate on detecting the third assumption

A

The third assumption is autocorrelation.

We use the Durbin Watson test

18
Q

elaborate on the Durbin Watson test

A

the null hypothesis is that the correlation between a residual and its lag-1 residual is 0. If the value we observe is extreme, we reject the null and say that there is likely autocorrelation here.

19
Q

downside of DW

A

only checks for lag-1

20
Q

how can we check for lag-l autocorrelation

A

use Breusch Godfrey

21
Q

elaborate on Breusch Godfrey

A

We regress the regular regression to find the residuals.

Then we create a new regression where the dependent variable is the residuals we just found, and as explanator variables we use the OG ones AND lagged variables up to some order k of our choice.

The R^2 of the new regression when multiplied by (T-r) (r is the k’th order we chose) follow a chi squared distribution with r degrees of freedom.

22
Q

what happens if assumption 4 is violated

A

Estimaror will not even be consistent

23
Q

how do we test for normality of residuals

A

Bj test, Bera Jarque

24
Q

elaborate on BJ test

A

based on checking whether the skewness and kurtosis holds up.

25
elaborate on testing for the wrong functional form
Ramsey's RESET test. The null hypothesis is that all alphas are zero. The R^2 is used.
26
downside of RESET test
Tells us nothinh on what the better form is. However, it will tell us if non linear is better
27
elaborate on what happens if we omit an important variable
If there is some correlation between the omitted variable and some of the remaining variables, their coefficeints will be biased. Even if there is 0 correlation, the constant term will be biased to account for the lacking information. This is not a good approach. The outcome is that we just simply get a worse-perofrming model that doesnt capture the relationship as well as it is possible to do.
28
elaborate on parameter stability testing
This is about figuring out if we have the ideal setup over the entire sequence of values etc. There are 2 approaches. 1) Chow (analysis of variance) 2) Predictive failure
29
elaborate on Chow test
The basic component is to perform regressions on the entire sample, and on sub samples. then acquire teh RSS of each. The test statistic is F-distributed and use a combinaiton of the RSS's. One can also use dummy variables for this test, where we add a dummy variable per variable to have, whcih indicates that the data point is present in this subsample. In this version, the coefficeints that correspoind ot the variables that also have the dummys are in the null hypothesis as all 0.
30
weakness of chow test?
requires a lot of data because we are doing 2 regressions
31