Week 4 (Multiple Regression) Flashcards

(39 cards)

1
Q

What is a regression?

A

Extends upon a correlation (relationship between two variables)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Multiple Regression

A

-Extension of simple linear regression
-Explores impact of multiple predictor variables
-Tests relationship in parallel

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Regression VS ANOVA

A

Regression
-Focuses on relationships between predictor variables and one outcome variable

Factorial ANOVA
-Focuses on differences in scores on the dependent variable, according to two or more independent variables.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Regression VS ANOVA (Requirements)

A

Regression
-Predictor variables can be continuous, ordinal or binary data, outcomes must be continuous.
-One hypothesis per predictor

Factorial ANOVA
-IVs must be categorical data, can have 2+ conditions, dependent variable must be continuous.
-One hypothesis per IV (=2) and one hypothesis for the interaction (so 3 in total)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Types of multiple regression

A

-Forced Entry
-Hierarchical Multiple Regression
-Stepwise Multiple Regression

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Forced Entry Multiple Regression

A

-Predictors based on previous research and theory
-Do not state a particular order for the variables to be entered
-All variables are forced into the model at the same time
-Known as Enter method in SPSS

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Hierarchical Regression

A

-Predictors based on previous research
-Researcher designs the order in which predictors entered into model
-Enter known predictors first and then enter new predictors

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Stepwise regression

A

-Based on maths rather than previous research/theory
-Both forward and backward methods
-Computer programme selects the predictor that best predicts the outcome and enters that into the model first

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Assumptions of multiple regression

A
  1. Sample Size
  2. Variable Types
  3. Non-zero variance
  4. Independence
  5. Linearity
  6. (Lack of) Multicollinearity
  7. Homoscedasticity
  8. Independence Erros
  9. Normally Distributed Errors
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Variable types (Regression Assumption)

A

-All predictor variables should be quantitative
*Can be continuous, categorical, or ordinal
-Outcome variable must be quantitative and continuous

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Non-zero variance (Regression Assumption)

A

-Predictor variables should have a variance
-In other words, should not have a variance of zero

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Independence (Regression Assumption)

A

-All values of the outcome variable should be independent
-Each value of the outcome variable should be a separate entity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Linearity (Regression Assumption)

A

-Assume that the relationship between the predictor and outcome variable will be linear
-If analysis is run on a non-linear relationship, the model can be unreliable

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Sample Size of Regression (Regression Assumption)

A

More is better
-Field (2010) suggests you use the following equations to identify an appropriate size
50+8k where k = number of predictor variables
104+k

Or can use power analysis

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Multicollinearity (Regression Assumption)

A

-Strong correlation between predictor variables
*Perfect collinearity when you have a correlation of 1 between predictors
-Becomes difficult to interpret results
-Untrustworthy beta values
-Can’t identify individual importance of each predictor
-Limits size of r squared
-Threatens the validity of the model produced

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Identifying multicollinearity (Regression Assumption)

A

-VIF (Variance Inflation Factor)
*If the average VIF is substantially greater than 1then regression might be biased
*If largest VIF is greater than 10 there is definitely a problem

-Tolerance
*If tolerance is below 0.1 a serious problem
*If tolerance is below 0.2 a potential problem

17
Q

What are residuals (Regression Assumption)

A

Distances between regression line and individual data points

18
Q

Homoscedasticity (Regression Assumption)

A

-At each level of the predictor, the variance of the residuals should be constant

19
Q

Independent Errors (Regression Assumption)

A

-For any two observations (data points) the residual points should not correlate, they should be independent
-This can be indentified as an issue with the Durbin-Watson Test

20
Q

Normally Distributed errors (Regression Assumption)

A

-The residual values in the regression model are random and normally distributed, with a mean of. I.e, there is an even chance of points lying above and below the best-fit line

21
Q

How to check sample size (up-front assumption)

A

-Calculate the desired sample size in advance

22
Q

How to check variable types (up-front assumption)

A

-Make sure your measures provides data appropriate for multiple regression

23
Q

How to check non-zero variance (up-front assumption)

A

Calculate the standard deviation of your variables, check if they have variance > 0

24
Q

How to check independence (up-front assumption)

A

-A measurement issue (make sure your outcome scores are all from different people)
-Should not have two or more values on your outcome variable from the same person

25
How to check linearity (up-front assumption)
Check by analysing residuals in SPSS
26
How to check lack of multicollinearity (up-front assumption)
Check VIF and Tolerance statistics in SPSS
27
How to check for Homoscedasticity (up-front assumption)
Check by analysing residuals in SPSS
28
How to check for independent errors (up-front assumption)
Check Durbin-Watson in SPSS
29
How to check for normally distributed errors (up-front assumption)
Check by analysing residuals in SPSS
30
Participants needed for different effect sizes (2 predictors)
Small effect: 478 Medium effect: 67 Large effect 31
31
Participants needed for different effect sizes (3 predictors)
Small effect: 543 Medium effect: 76 Large effect: 36
32
Participants needed for different effect sizes (4 predictors)
Small effect: 597 Medium effect: 84 Large effect: 39
33
Participants needed for different effect sizes (5 predictors)
Small effect: 643 Medium effect: 91 Large effect: 43
34
R squared value in regression
How good is our model at explaining our data variance explained by the model/ variance not explained by the model effect/error E.g, R squared = 0.887, so 88.7%
35
Interpret Durbin Watson Test
-Tests whether residuals next to each other are correlated -Varies between 0 and 4 -2 means residuals are uncorrelated -Value greater than 2 indicates a positive correlation -Values greater than 3 or less than 1 indicate a definite problem -Values close to 2 suggest there is no issue
36
How does f value ANOVA work
Mean square value 1 divided by value 2 = f value
37
Unstandardised and Standardised Beta
Unstandardised beta (b): The change in Y for a unit change in X Standardised beta (β): The change in Y for a standardised change in X
38
T-test
Which predictors individually predict our outcome
39
What is show on a coefficients table
-Unstandardised beta -Standardised beta -T-test -Collinearity (VIF and Tolerance)