Econometrics Flashcards

1
Q

TSS

A

ESS+RSS

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Degrees of freedom for TSS

A

n-1

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Degrees of freedom for ESS

A

k-1

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Degrees of freedom for RSS

A

n-k

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

MSS

A

XSS/df

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

R-Squared

A

ESS/TSS, The proportion of the total sample variation that is explained by the independant variable.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Adjusted R^2

A

1-(1-R^2)*((n-1)/(n-k)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

F test

A

(Explained SS/(df) )/(RSS/(df)), use table to find level of significance.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

t-test for significance of coeficients

A

(beta-0)/se

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

confidence interval

A

mean + or - se*t, t using the degrees of freedom n-1

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Population Regression Function

A

E(y*bar**x), it is a linear function of x. It is constructed from the population data. How the conditional expecatation of y changes with x.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

OLS regression analysis

A

The technique of estimating the coefficient of population regression by minimising the sum of the square residuals between the sample regression function and the observed data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Dummy Variable

A

a binary variable where the observations are either 0 or 1. Usually describes different groups, or states of nature. aka qualitative variables. Where x=0 is the baseline case and x=1 is non-baseline case.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Chow test

A

restricted F-Test for structural stability. Used to test the stability of the coefficients when the data is split at a specific point. Usually used in time-series data to test for the presence of a structural break. This is done by comparing the restricted model to the unrestricted model.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Goodness of Fit

A

it is defined as ESS/TSS. This is a measure of the proportion of the total variation in Y that is explined by the regression model. And as such gives the goodness of fit. Since TSS = ESS +RSS, SUM(Yi-Ymean)^2 = SUM(Yhat-Ymean)^2+Sum(ui)^2. Do not exceed 1 or less than 0

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Assunption 1

A

linear in perameters

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

assumption 2

A

random sampling

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

assumption 3

A

sample variation in explanatory variables

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

assumption 4

A

0 conditionalmean. error term has an EV of zero.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

Autocorrelation

A

correlation between the errors in different time periods

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

Bias

A

The difference between the expected value of an estimator and the population value of of the value.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

critical value

A

in hypothesis testing, the alue against which a test statistic is compared to determine whether or not the null hypothesis is rejected.

23
Q

Fitted value

A

the estimated values of the dependant variable, when the regression is run.

24
Q

heteroskedasticity

A

The variance of the error term is not constant.

25
Q

kurtosis

A

the measurement of the the thickness of the tails of the distribution

26
Q

least squared estimator

A

an estimator that minimises the sum of the squared residuals.

27
Q

Multicolinearity

A

correlation among independant variables

28
Q

Power of a test

A

the probability of rejecting the null hypothesis when it is flase. Depends on the values of the population perameters under the alternative.

29
Q

residual

A

the difference between the actual value and the fitted value. there is a residual for each observation.

30
Q

testing the marginal contribution of variables

A

h0: restriction is true, e.g. B3 = 0, h1 restriction is not true. (RSSr - RSSur/m)/(RSSur/n-k)
where m is the number of restrictions, n is the number of observations, k is the number of perameters in the unrestricted case. This is an F-Test is F is greater than the critical value then reject the null. ‘variables provide a significant marginal contribution.’

31
Q

F test restiction R^2 form

A

(R^2ur-R^2r/m)/(1-R^2ur/n-k)

32
Q

Multiple regression assumption 5

A

Homoskedasticity.

33
Q

Why does heteroskedasticity cause problems with OLS

A

Since we have to estimate the variances of the perameters in our regression, we use the homoskedasticity assumption. And our standard errors are based entirely on this, therefore t-test may not work.

34
Q

Causes of heteroskedasticity is cross-sectional

A

Error variance may increase proportionally with a variable in the model. Since cross sectional data often includes very small and very large values, the variance may be higher at higher values of x.

35
Q

consequences of heterosedasticity

A

perameters are still unbiased. the variance of betas is different under heteroskedasticity, (efficiency). P tests and t tests will be wrong as the homoskedasticity assumption is used to estimate s.e.

36
Q

GLS

A

Transforming the model to create a new model tha obeys the classical assumptions. Since the variance is not constant we can represent it as w^2*o^2. By dividing all of the perameters by w, we can find a new error term. o/w. 1/w^2 * w^2 *o^2 = o. All variables are now divided by w. In GLS, it now obeys classical assumptions so it is unbiased and will have a smaller variance than OLS on the original model.

37
Q

Park test for heteroskedasticity

A

find the squared residuals from the original regression. Run the regression. lnu^2 = α + βln(Xi) + Vi. We can then test Ho: β = 0. Significant β suggests heteroskedasticity is present. However, this does not rule out that the error term in the new regression is not heteroskedastic itself. and therefore the significance of the beta could be wrong.

38
Q

Goldfields-Quandt Test for heteroskedasticity

A

We estimate the varince of Ui as the RSS/n-k. Order the data by magnitude of Z (explanatory variable), omit some observations in the middle. Run two seperate regression, with low values and high values. Find the RSS for both. the test statistic is
o2^2/O1^1 = (RSS2/n2-k)/(RSS1/(n1-k). If this test statistic is greater than the critical F value then we can rejcet the null hypothesis that the variance is constant. Only tests one variable.

39
Q

White Test for heteroskedasticity

A

Run OLS on the original regression, find residuals. Run a second regression with Ui^2 as the dependant variable and the original variables, variables squared and interaction effects as the dependant varaibles. Obtain R^2 from this regression. Calculate the White test statistic. W = nR^2. This test statistic has a chi^2(p) distribution where p is the number of regressors in the auxillary regressiion (not the constant.)

40
Q

Breusch-Pagan-Godfrey Test for heteroskedasticity

A

Run original regression and obtain residuals. Use this to construct a varibale Pi. Pi = (ui^2)/Sum(U^2i/n). Run a second regression with p as the dependant variable. Calculate BPG = ESS/2. On a Chi^2(p) distribution.

41
Q

Correcting for Heterskedasticity if the error variances vary directly with an explanatory variable.

A

GLS. So wi = Xi. Divide all variables by Xi. Rerun regression

42
Q

Correcting for heteroskedasticity if the error variances vary inversely with an explanatory variable.

A

GLS. divide by 1/Z^2. Issues: if Z=0 doesnt work, dummy variables etc.

43
Q

If we cannot do GLS

A

Use OLS but with the corrected formula for variance. Using E(u^2) as an estimate for o^2w^2. Then use this estimated variance for intervals. Pros: easy to calculate Cons: not as efficient.

44
Q

Types of incorrectly specified Model

A

Omission of relevant Variables, Inclusion of irrelevant variables, measurement error.

45
Q

Issues with incorrect Specification

A

the Expected value of Beta doesnt equal the population value., we obtain a biased estimate of beta 1.

46
Q

RESET test for model specification

A

Rerun a regression with the dependant variable squared and cubed as explanatory varibles. ho: betay^2 = betay^3 = 0, if it is rejected then model is correctly specified. Do normal restricted f-test where UR is the regression with added Y^2 etc.

47
Q

Lagrange multiplier test for model specification

A

obtain Ui from an estimated restricted regression. Run auxiliary regression: regress residuals on all regressiors including omitted ones. LM = nR^2 Chi^2 distribution. if it is significant reject restricted regression.

48
Q

autocorrelation

A

When the assumption that the errors are uncorrelated with one another is violated.

49
Q

interpretation of beta in a linear regression

A

beta = dy/dx

50
Q

Interpretartion of beta in log-linear model

A

beta = dlny/dlnx, not a marginal effect. this is an elasticity. dy/dx * x/y.

51
Q

interpretation of beta in semi-log model

A

B = (dy/y)/dx for lny = … or B = (dy)/(dx/x) for y= b + bln(x).

52
Q

estimated variance of Beta

A

sigma^2*SUM(X^2)/(SUM(X^2))^2

53
Q

Expected value of beta

A

b + (1/SUM(x^2))*SUM(XiE(ui))