Week 7 - SMR/HMR Flashcards
(43 cards)
• List the two types of tests in multiple regression (MR)
Standard
Hierarchical
• Explain the different questions that are addressed in bivariate regression (x1)
o Does the predictor account for significant variance in the DV?
• Explain the different questions that are addressed in multiple regression (x5)
Extending on bivariate regression
Do predictors jointly account for significant variance in the DV?
• F test of Model R2
For each IV: does it uniquely account for variance in the DV?
• t-test of β for each IV
• Explain (in words) what R2 represents in each of the following cases:
o Bivariate regression
o Multiple regression with uncorrelated predictors
o Multiple regression with correlated predictors
(x2)
• Strength of overall relationship between criterion and predictor/s
Variance accounted for by the IV/set of IVs
What is r2 in bivariate regression? (x3)
coefficient of determination -
Proportion of variance in one variable that is explained by variance in another
(like eta2 in ANOVA - SSeffect/SStotal)
What is error variance in bivariate regression? (x2)
1 -r2
SSresidual/SSy
• Conceptually speaking (i.e. in words), how is shared variance determined in:
o Multiple regression with uncorrelated predictors
(x1)
There isn’t any,
So R2 is just r2 for each IV/predictor added together
• Conceptually speaking (i.e. in words), how is shared variance determined in:
o Multiple regression with correlated predictors (x3)
It’s the overlap of the predictors with each other
So you have to account for it - so it doesn’t get used twice in calculating R2 (which measures non-redundant variance)
(can’t just add, as you do for uncorrelated)
• Explain the difference between a zero-order (Pearson’s) correlation, a partial correlation and a semi-partial correlation (x1, x1, x1)
Zero-order - only 1 source of variance in DV, so none shared
Partial: relationship between predictor 1 and the criterion, with the variance shared with predictor 2 partialled out of BOTH dv and iv
Semi-partial: relationship between predictor 1 and criterion, after partialling out of predictor 1 variance shared between predictor 1 and 2
• In a Venn diagram representing one criterion and three predictors, indicate how shared variance between predictors 1 and 2 would be represented, and how the unique variance of predictor 3 would be represented. (x2)
1 and 2 would overlap with each other, and the DV
While 3 would only overlap the DV
• List the 4 key differences between the structure of ANOVA tests and MR tests
No test of overall model in ANOVA - auto in MR
Main effect of IV in ANOVA disregards effect of other variables - MR tests unique variance in each (controls for other variables)
Interactions auto in ANOVA - hard in MR (need MMR)
ANOVA = Fs, effect sizes for IVs/interaction, plus follow-ups
*MR = Rs F-test, beta t-tests, plus follow-ups
What is semi-partial correlation squared (sr2)? (x3)
proportion of variance in DV uniquely accounted for by IV1, out of total
- A/ A + B + C + D
- (where C and D are shared IV2/DV variance)
What is partial correlations squared (pr2)? (x3)
Proportion of residual variability in DV accounted for by IV1, after IV2 variance removed
- A/A + B
- (where a is unique, b is unaccounted for DV variance)
• Explain the difference between the semi-partial correlation squared (spr2/sr2) and the partial correlation squared (pr2)
Partial is like partial eta2 - leftover variability in DV that IV accounts for
Semi-Partial is like eta2 - bit of total DV variability that uniquely due to IV
*the go to effect size for regression
• Identify the linear model for a multiple regression analysis with 2 predictors (x2),
And explain what b1, b2, and a represent
Ŷ = b1X1 + b2X2 + a
Ŷ - predicted score is still a function of slopes, X scores and constant
b1 - slope of plane relative to x1-axis
b2 - slope of plane relative to x2-axis
a - the constant (y-intercept, when x1 and x2 = 0)
• Explain why means and standard deviations are not as critical for interpreting MR results as they are for t-tests and ANOVAs
Because you’re not interpreting/comparing means, but the direction relationship between variables
Although SD still tells us change in DV for every SD change in it’s IV
• Explain what Cronbach’s is (x2), and what values of this index we would ideally like to have (x2)
index of internal consistency (reliability) for a continuous scale
*how well items “hang together”
Use scales with high reliability ( > .70) if available – less error variance
• Define validities (x1), and identify their ideal levels (i.e. high or low) (x1)
Relationship between IV and DV
High - show a strong association, yay!
• Define collinearities (x1), and identify their ideal levels (i.e. high or low) (x1)
Relationships between IVs
Low, because the higher they are, more chance of redundant IV
• Explain the principle of parsimony (x2)
The simplest explanation for the data is good
ie predictors explain different variance in the model
How does parsimony relate to validities and collinearities (x2)
More parsimonious when high validity (predictor/criterion relationships)
And low collinearity (predictor relationships)
• Define R (x1) R2, (x1) Radjusted, and R2adjusted (x1) in the context of multiple regression
Multiple correlation coefficient (R) is bivariate correlation between criterion (Y) and best linear combination of the predictors (Ŷ)
R2 - square R to find variance in Y accounted for by composite (Ŷ)
The adjusted versions attempt to correct positive sample bias by correcting for sample size (so no point for 30+)
• Identify the test used to assess the overall variance explained (x1, plus explain x1), and explain what a significant result means (x1)
F-test: MSregression/MSerror
*variance accounted for (R2/p) divided by that not accounted for (error/N - p - 1)
The relationship between predictors (as a group) and criterion is different from zero
• If the sr values for the predictors are known, explain how to work out the unique variance (x1) explained by each predictor and the variance shared between all predictors (x1)
Square the semi-partial correlation to get the coefficient
Shared variance = R2 - sum of all sr2