M9 - Simple Linear regression Flashcards

1
Q

Population vs sample

What do we measure?
Why?

A

We measure the info that is contained in the data.

Measure only x and y
U is not observable

–> if it would be, we could determine the coeff exactly

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

How do we measure b0 and b1?

A

We only have estimated values of b0 and b1

–> determine the estimators

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Whats OLS?

A

Ordinary Least Squares

–> choosing both coeff in a way that the sum of the squared residuals is minimized.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Estimator of the slope coeff

Only correct if….

A

b1

Only correct if denominator is +

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

OLS summary

  • the … of the slope coeff …. = …/…
  • if x and y are + …., the slope coeff is …
  • if x and y are - …., the slope coeff is …
  • x needs + ….. to determine an estimator
A

OLS summary

  • the ESTIMATOR of the slope coeff B1 = COV(x,y) / VAR(x)
  • if x and y are + CORRELATED the slope coeff is POSITIVE
  • if x and y are - CORRELATED the slope coeff is NEGATIVE
  • x needs + VARIANCE to determine an estimator
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Is it better to have more of the structural or the stochastic term?

A

The more of the variables in the structral term the better

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Variance decomposition

A

Total variance: actual yi - mean y

Explained v: predicted yi - mean y

Residual v : actual yi - predicted yi

Total = explained + residual

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

R square?

A

How well does the empirically tested model fot the data?

R-square = 1 - total var/residual var

Interval [0,1] thehigher the better

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Assumptions of the simole regression model?

Needed
1-6

Whaz happens if not fulfilled?

A
  1. the linear model describes the relship between x &y
  2. sampe through random draw of the pop
  3. the expected value for the disturbance term is 0
  4. the iv is not constant
  5. u not corr with x
  6. x values have been measured accurately

If not fulfilled, results are biased

  1. Homocedasticity: u has constant variance, which ins not correlated with x
  2. error terms ui are normally distributed
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Assumptions of simple regression model

Not needed
7-8

A
  1. homoscedasticity
    Disturbance term u has const variance, which is not corr with x
  2. error term u is normally distributed
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Heteroscedasticity

effect

A

differing variance across all values of an IV

e.g. age with income

  • -> OLS is biased and inefficient
  • -> SE are distorted
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Homoscedasticity

A

variance is the same across all values of IV

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Dealing with heteroscedasticity

-
Solution

A

+ ols estimates are UNBIASED

  • ols estimates are UNEFFICIENT
  • SE lf the coeff estimates by OLS are DISTORTED

Solution : use robust estimates

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Testing for significance of the coeff?

Is there a …. between x and y?
H0: is it ….. 0 ?

Distribution of predicted b1:
In large samples: b1 …. to a …. distribution
In small samples: b1 is ….. distributed

A

Testing for significance of the coeff?

Is there a RELATIONSHIP between x and y?
H0: is it UNEQUAL 0 ?

Distribution of predicted b1:
In large samples: b1 CONVERGES to a NORMAL distribution
In small samples: b1 is NORMALLY distributed

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Multiple regression

  • at least 2 … variables
  • b0: measures of the …
  • b1:
    -b2:
    B1+ B2 –> … …

“How …is the … size of x1, provided that x2 remains constant?”

A

Multiple regression

  • at least TWO EXPLANATORY variables
  • b0: measures the INTERCEPT
  • b1: SLOPE of the linear relship x1&y
    -b2: SLOPE of the linear relship x2&y
    B1+ B2 –> EFFECT SIZE

“How LARGE is the EFFECT size of x1, provided that x2 remains constant?”

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Problems in multiple regression?

A
  • ignored variables –> omitted variable bias
  • relship iv and dv?
  • relship between ivs? –> multi-collinearity
  • too many regressors?
  • wrong functional term?