Reliability and validity Flashcards

(21 cards)

1
Q

Test-retest reliability

A

Consistency across time

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Internal consistency

A

Consistency across items

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Inter-rater reliability

A

Consistency across researchers

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is used to assess test-retest reliability

A

ICC

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is used to measure internal consistency

A

Chronbach’s alpha or McDonald’s omega

Mcdonald’s omega is the preferred option

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What are they advantages of Mcdonald’s omega over Cronbach’s alpha

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Validity

A

The extent to which the scores actually represent the variable they are intended to

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Which type of ICC do we use when writing up the output

A

The ICC2

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

ICC

A

Intraclass correlations - look at the absolute agreement between variables

Close to 1 - good - most variation comes from difference between individuals

Close to 0 - bad - most variation comes from measurement error or differences within individuals

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Values for ICC

A

0.5 to 0.75 = moderate reliability

0.75 to 0.9 = good reliability

> 0.9 = excellent reliability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Internal consistency

A

The consistency of people’s responses across the items on a multiple-item measure

In general, all items on a questionnaire should reflect the same underlying construct. Therefore, people’s scores on a set of items should be correlated with each other

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Split-half method

A

Can be used to assess internal consistency

This method involves splitting the items on a questionnaire into two halves with each half measuring the same elements but in slightly different ways.

Then a score is computed for each set of items and the relationship between the two sets of scores is examined

If a scale is very reliable, a person’s score on half of the scale should be the same or similar to their score on the other half.

A split half correlation of +.80 is generally considered good internal consistency

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What is the problem with the split-half method

A

There are several ways in which the set of data can be split into two and so the results could be a product of the way in which the data were split

e.g. someone compares the first half of the questionnaire to the second half and finds out the internal consistency is bad so they decide to do odds and evens comparison and find out its good - they publish that and then everyone uses that questionnaire - flexibility can be an issue

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What is the most common measure of internal consistency

A

Cronbach’s a

Cronbach’s alpha - the extent to which different items on the same test correlate with each other

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Alpha coefficients

A

Range from 0 to 1

The higher the score, the more reliable the scale is

A value of +.70 is generally taken to indicate good internal consistency

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Problems with Cronbach’s alpha

A

It’s a lower bound estimate - meaning it gives the lowest estimate of reliability (it’s pessimistic)

Assumption of tau-equivalence

More questions = higher alpha

17
Q

Tau-equivalence

A

The assumption that all items have the same factor or component loadings - this is unlikely and can reduce alpha estimates by up to 11%

18
Q

McDonald’s Omega (w)

A

Works in very much the same way as Cronbach’s alpha but does not require tau equivalence, so it works even when items vary in their contribution to the total score

19
Q

Omega Hierarchal

A

Assesses the extent to which variance on a measure is due to a general factor (g)

For example, an intelligence measure may have discrete factors (spatial intelligence, emotional intelligence etc) but should also tap into a general factor (intelligence)

20
Q

Omega total

A

Assesses reliability for all factors (general and other factors)

21
Q

How is inter-rater reliability assessed

A

Using Cronbach’s alpha and ICC’s when the judgements are quantitative or Cohen’s k when the judgements are categorical