Model evaluation and violations Flashcards

1
Q

Model evaluation:

A

“Models are devices that connect theories to data. A model is an instantiation of a theory […]” (Rouder et al., 2016, p. 2)
– Violations of model assumptions reduce reliability.
– Possibility leading to incorrect conclusions about data.
– Violations don’t render model results wrong; neither are models without violations necessarily correct.
– Generally we need to be cautious with and conscious of violations when interpreting results.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Assumptions of parametric models

A
–	e.g. t-test, ANOVA, linear regression
–	Remember LINE
–	Linearity (for continuous predictor variables)
–	Independence
–	Normality
–	Equal of variance (also homogeneity)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Using residuals to evaluate models:

A

– Linearity: linear relationship between outcome and predictor(s)
– Independence of residuals (remember iid)
– Normality of residuals
– Equality of variance of residuals

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What are residuals?

A

Residuals are the unexplained (residual) variance: error in the modelling results.
Distance between observed (y) and predicted rt (y ̂): ϵ=y-y ̂
The closer the residuals are to 0, the lower the prediction error.

– Distributed around 0
– Right / positive skew shows some violation of normality assumption
– rts are 0 bound and known to have a heavy right tail (see e.g. Baayen, 2008)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What can we do about positive skew:

A

– Logarithmic transformation is routinely used in the literature, especially for rts, to correct positive skew (see e.g. Baayen, 2008).
Can also be used for non-linear relationship.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Linear / arithmetic scale

A

Distances between adjacent values must be the same, i.e. ±1 on a linear scale.
Distances between 1 inch and 2 inch is 1 etc.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Logarithmic scale

A
Distance between units is going down as values go up.
	Granularity for small while also including large numbers.
	Basis 10 (⋅ or ÷ by 10): 0.1, 1, 10, 100
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

skewness vs kurtosis

A

– Skewness is about horizontal shape of the distribution
– Kurtosis is about vertical shape of the distribution

Normality: Kurtosis= the sharpness of the peak of a frequency-distribution curve.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Independence of residuals

A

– Plot residuals across ppt id, predictor, predicted data
– What are we looking for?
– Dependencies:
• linearity, independence violations
• For any value on the x-axis, mean of residuals should be roughly 0.
– Equality of variance:
• For any value on the x-axis, the spread of the residuals should be about the same (constant variance).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Equality of variance

A

Homogeneity (same) vs heterogeneity (different)
Variance = SD^2: deviations between observations and their mean
More diverse groups have a larger variance

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Linearity

A

– Linear relationship between outcome and predictor variable
– Relationship can be described with a straight line
– Increase is additive.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Linearity: continuous predictor

A

– Linearity: mean of residuals should be 0 at any age.
– Divided age into 20 equal sized bins.
– Calculated the mean of the residuals for each bin.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly