Lecture 3 - Quantitative Methods Flashcards

1
Q

Difference b/w chi-square and regression analysis

A

Þ Chi-square analysis looks at the association between two categorical
variables.
Þ Regression analysis looks at predicting one continuous variable from another

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is a correlation?

A

the association b/w 2 continuous variables

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Correlation is a form of. . .

A

BIVARIATE analysis (measure of relationship b/w 2 variables)

Correlation quantifies the relationship (typically linear) between two variables X & Y
in terms of direction, degree.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

An association b/w 2 variables can be. . .

A

LINEAR or NON-LINEAR

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What does the Pearson correlation coefficient (r)?

A

measures the linear association between two
continuous variables.

It compares how much the two variables vary together to how much they vary separately.

It ranges from -1 to 1. A r value of 0 indicates no association between 2 variables.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

R squared (r^2) is. . .

A

the percentage of variance accounted for (i.e. how much of the variation in the outcome variable is explained by the predictor)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is variability?

A

how much a given variable varies from observation to observation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is covariability?

A

how much two variables vary together

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Degree of relationship

A

® Small effect: 0.1 < r < 0.3, or -0.3 > r > -0.1

® Medium effect: 0.3 < r < 0.5, or -.3 > r > -0.5

® Large effect: 0.5 < r < 0.7, or -0.5 > r > -0.7

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What can greatly influence the value of a correlation?

A

extreme scores / outliers

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Pearson’s r is. . .

A

BIDIRECTIONAL

  • meaning, the correlation b/w Variable B and Variable A is the same as that b/w Variable A & B
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Regression toward the mean:

A

® With imperfect correlation, an extreme score on one measure tends to be followed
by a less extreme score on the other measure.

® This is because extreme scores are often due to chance. And if it’s due to chance, it’s
extremely unlikely that the other value will also be extreme.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Null hypothesis testing

A

we assume there is no effect (i.e. no association)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

The null hypothesis –

A

ρ = 0. Rho (ρ) is the correlation in the population

® If the probability of finding an r this big if the real association in the population (ρ) is small (p < .05), we reject the null hypothesis. So, we infer that the correlation value for the population is NOT zero. There is significant association between the two
variables.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

ρ will depend on. . .

A

the size of the sample (n)

it will also depend on
the degrees of freedom (df).

® For correlation, df = n – 2.
® For a small correlation to be significant, a high df (lots of participants/data) is
required.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

When is Spearman’s correlation coefficient used?

A
  • when the data is ORDINAL (ranked) & when the data is one-directional but NON-LINEAR
  • Convert the data to ranks before calculating correlations. Converting to ranks can linearize non-linear data.
  • Useful for non-linear data and data sets with outliers.
17
Q

What is regression used for?

A
  • predicting one variable from another
  • However, unless the correlation is perfect (r = +1 or r = -1), the prediction will not be
    exact.
  • Looks for the line of best fit (regression line)
18
Q

Errors in regression are assumed to be

A
  • independent
  • normally distributed (w a mean of 0)
  • homoscedastic (equal error variance for levels of predicted Y)
19
Q

How is the assumption of the independence of errors assessed in regression?

A
  • by reflecting on the sampling procedure for the study (and NOT by looking at plots of residuals)
20
Q

What does the least squares parameter estimate?

A

estimate parameters by minimizing the total squared error (regression coefficients)

21
Q

What does the ANOVA (F test) for aggression tell us?

A
  • F test tells us whether the variance explained is significantly different from zero
  • if the F test is NOT significant, the regression is worthless. The predictor does not explain the outcome variable at all

® Null hypothesis: r^2 = 0.
® The F test has 2 degrees of freedom. For simple regression, the first is always 1, the
second df is the same as for the earlier t-test of regression coefficients.

22
Q

Regression parameters (a, b)

A

® The intercept (a) is the estimated value of Y when X = 0.
® Slope (b) (gradient) (the standardized/regression coefficient in JASP) indicates
whether there is a relationship between X & Y, whether that relationship is positive
or negative, and the estimated change in Y when X increases by 1. For H0: b = 0
® Slope can be transformed into standardized form (covert X & Y into z-scores, and
then do the regression). This is called a standardized regression coefficient (called
beta in SPSS/JASP), and is the same as the correlation coefficient for bivariate data.

23
Q

Regression diagnostics

A

a) Histogram of residuals – normality: we want the errors to be approximately
normally distributed

b) Residuals plot – homoscedasticity: a scatterplot of residuals against predicted values
to check for heteroscedasticity. Absence of any systemic pattern supports the
assumption of homoscedasticity (where the variance of the residual/error term in a
regression model is constant).