Part 7. Intro to Linear Regression Flashcards

1
Q

Regression analysis

A

A tool for examining whether a variable is useful for explaining another variable.

i.e. whether earnings growth/cash flow growth helps explain the company’s value in marketplace.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Sum of squares total (SST)

A
  • A simpler exploration to try understand why each company’s ROA differs from mean ROA of 12.5%, we look at the sum of squared deviations of obs from mean to capture variations in return on assets (ROA) from their mean.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Simple linear regression (SLR)

A

Method for relating dependent and independent variables through estimation of a relationship, where we have one independent variable.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Multiple regression

A

Method for relating dependent and independent variables through estimation of a relationship, where we have more than one independent variable.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Ordinary least squares (OLS) regression

A

The goal is to fit a line to observations on Y and X to minimise the squared deviations from line; using least squares criterion.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Line of best fit

A

In simple linear regression, the estimate intercept, b0^, and slope b1^ are such that the sum of squared vertical distances from the observations to fitted line is minimised.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Residual for ith observation, ei

A

This is how much the observed value of Yi differs from the Yi^ estimated using the regression line: ei = Yi - Yi^.

This refers to the true underlying population relationship, whereas the residual refers to fitted linear relation based on sample.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Residuals

A

Represented by the vertical distances from the fitted line, therefore in the units of measurement represented by the dependent variable.

i.e. if dependent variable is in euros, the error term is in euros etc.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Sample correlation, r

A

The ratio of the covariance to the product of the standard deviations.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Slope

A

The change in dependent variable for one unit change in independent variable.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Cross sectional regression

A

This involves many observations of X and Y for same time period, depending on regression model these observations could come from different companies, asset classes, investment funds etc.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Time series

A

Use many observations from different time periods for the same company, asset class, investment fund, country or other entity depending on regression model.

i.e. monthly data from many years to test whether country’s inflation rate determine short term interest rates.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Assumptions of simple linear regression model:

A
  1. Linearity - the relationship between the dependent variable Y, and independent variable X is linear.
  2. Homoscedasticity - the variance of regression residuals is the same for all observations.
  3. Independence - the observations, pairs of Ys and Xs are independent of one another, implies regression residuals are uncorrelated across observations.
  4. Normality - the regression residuals are normally distributed.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Homoskedasticity

A

The variance of the residuals is the same for all observations.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Heteroskedasticity

A

If residuals are not homoscedastic, if the variance of residuals differs across observations.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Sum of squares regression (SSR):

A

The sum of the squared differences between the predicted value of the dependent variable, Yi^, based on the estimated regression line and mean of dependent variable Y-.

17
Q

Coefficient of determination (R^2):

A

The percentage of the variation of the dependent variable that is explained by the independent variable.

  • measure used to evaluate goodness of fit.
18
Q

Standard error of estimate (se):

A

The absolute measure of distance between the observed values of the dependent variable, and those predicted from estimated regression.

The smaller the se, the better fit of the model.

19
Q

Measure of goodness fit of estimated regression:

A
  1. Standard error estimate (se)

2. F-statistic

20
Q

Standard error of slope coefficient (sb1)

A

In simple linear regression, this is the ratio of the models standard error of the estimate (se) to the square root of the variation of the independent variable.

21
Q

Indicator/dummy variable

A

The case where it takes only the values 0 or 1 as the independent variable.

22
Q

Level of significance

A

Its always a matter of judgement.

  • There is a 5% chance of rejecting H0, which is true (Type I error).
  • Decreasing level of significance from 0.05 to 0.01 decreases the probability of Type 1 error, but also increases probability of Type 2 error - failing to reject H0, when it is false.
23
Q

p-value

A

The smallest level of significance at which H0 can be rejected.

  • The smaller the p-value, the smaller the chance of Type 1 error (rejecting true H0), so greater likelihood the regression model is valid.
24
Q

The standard error of forecast depends on:

A
  1. the standard error of the estimate, se.
  2. the number of observations, n.
  3. the forecasted value of independent variable, Xf, used to predict dependent variable and its deviation from estimated mean, X-
  4. the variation of independent variable.