W8&9 - Multiple Linear Regression Flashcards

1
Q

What does simple linear regression do?

A

Quantifies the variance (R^2) in the DV that can be explained by the variance in the IV.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What does multiple linear regression test?

A

What combination from several IV explains the variance in the DV.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Can IV overlap in multiple linear regression?

A

YES

They can correlated with one another to explain the variance in the DV

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is unexplained variance known as especially when looking at a graph?

A

Residuals

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

How many people should you have per variable in the final model for multiple linear regression and why?

A

At least 10 per variable

Otherwise the variable coefficient can become unreliable.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What is the forced entry approach on SPSS in regards to multiple linear regression?

A

Means variables can still be in the model even if they’re not significant.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

List the types of multiple regression model building processes

A

Stepwise/forward

Hierarchical

Forced entry

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Define stepwise

A

Data driven + SPSS selects which variables are entered

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Define hierarchical

A

Researcher decides the order in which variables are entered

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Define forced entry

A

All predictors are entered into 1 model simultaneously

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Which out of the types of multiple regression model building processes doesn’t determine the unique variance that each IV adds to the model?

A

Forced entry

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What does the stepwise identify?

A

IV that explains the most variance in the DV + puts it in step 1 of the model.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What does the stepwise do once it has identified the IV that explains most the variance in the DV?

A

Looks for the IV that explains the most of the remaining unexplained variance + is included in the model provided it explains a SIG amount of the remaining variance.

This is repeated until there are no IVs left that explain further variance w/ a p<0.05.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

When is stepwise used?

A

When data driven

Not a theory drive

And if there are too many variables that you couldn’t possible know which ones are most predictive.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Define partial correlation

A

How strongly each of the variables correlate with the remaining information once we have explained or removed the variation one can explain.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What does a higher partial correlation do to the p value?

A

Makes it smaller

17
Q

Hierarchical regression analysis

A

Researcher decided order in which variables are enter

  • Order should be based on previous research or a plausible theory.
18
Q

What is normally entered in step 1 of the hierarchical regression analysis?

A

Known confounders

i.e age, gender + ethnicity

19
Q

What is normally entered in step 2 of the hierarchical regression analysis?

A

Known predictors

20
Q

What is normally entered in step 3 of the hierarchical regression analysis?

A

Test variables

21
Q

When would a hierarchical regression analysis be used over stepwise or forced entry?

A

When a ‘new’ variable of interest needs to be tested as for whether it explains further variation in a DV

22
Q

List the assumptions for multiple regression analysis

A

No multi-collinearity between predictors

Homoscedasticity of residuals

Linearity of residuals

Normality of residuals

23
Q

When does multi-collinearity occur?

A

When at least 2 predictors are highly correlated w/ each other in the final model.

24
Q

What can multi-collinearity lead to?

A

Unreliable regression b-coefficients for the predictors.

25
Q

What does the VIF (variance inflation factor) tell us?

A

How much the SE of the b-coefficient have been inflated.

Don’t want this as it widens the CI meaning you have less chance of showing that coefficient to be stat sig from 0.

26
Q

VIF>5 (r>0.90)

A

Maybe re-run regression after removing 1 of the ‘highly correlated’ variables (the least sig)

27
Q

VIF = 3.33

A

Assume there is no or low multi-collinearity