Chapter 4: Reliability Flashcards

1
Q

Reliability

A

Consistency or stability of test scores

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Factors that impact reliability

A
When the test is administered
Items selected to be included
External distractions (ex- noise)
Internal distractions (ex- fatigue)
Person administering test
Person scoring test
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Two components of score

A
True score (representative of true knowledge or ability)
Error score
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Systematic error

A

Error resulting from receiving a different set of instructions for test

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Classical test theory equation

A

Xi=T+E
Xi- obtained score
T- true score
E- error

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What measurement error reduces

A

Usefulness of measurement
Generalizability of test results
Confidence in test results

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Content sampling error

A

Difference between sample of items on test and total domain of items

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

How good sampling affects error

A

Reduces it

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Largest source of measurement error

A

Content sampling error

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Time sampling error

A

Random fluctuations in performance over time

Can be due to examinee (fatigue, illness, anxiety, maturation) or due to environment (distractions, temperature)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Inter-rater differences

A

When scoring is subjective, different scorers may score answers differently

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Clerical errors

A

Adding up points incorrectly

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Reliability (mathematic definition)

A

Symbol: rxx

Ratio of true score variance to total score variance (number from 0 to 1, where 0 is total error and 1 is no error)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Reliability equation

A

rxx= (sigma^2T)/(sigma^2X)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Reliability’s relation to error

A

Greater the reliability, the less the error

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What reliability coefficients mean

A

rxx of 0.9: 90% of score variance is due to true score variance

17
Q

Test-retest reliability

A

Administer the same test on 2 occasions
Correlate the scores from both administrations
Sensitive to sampling error

18
Q

Things to consider surrounding test-retest reliability

A

Length of interval between testing
Activities during interval (distraction or not)
Carry-over effects from one test to next

19
Q

Alternate-form reliability

A

Develop two parallel forms of test
Administer both forms (simultaneously or delayed)
Correlate the scores of the different forms
Sensitive to content sampling error (simultaneous and delayed) and time sampling error (delayed only)

20
Q

Things to consider surrounding alternate-form reliability

A

Few tests have alternate forms

Reduction of carry-over effects

21
Q

Split-half reliability

A

Administer the test
Divide it into 2 equivalent halves
Correlate the scores for the half tests
Sensitive to content sampling error

22
Q

Things to consider surrounding split-half reliability

A

Only 1 administration (no time sampling error)
How to split test up
Short tests have worse reliability

23
Q

Kuder-Richardson and coefficient (Cronbach’s) alpha

A

Administer test
Compare each item to all other items
Use KR-20 for dichotomous answers and Cronbach’s alpha for any type of variable
Sensitive to content sampling error and item heterogeneity
Measures internal consistency

24
Q

Inter-rater reliability

A

Administer test
2 individuals score test
Calculate agreement between scores
Sensitive to differences between raters

25
Composite scores
Scores that are combined to form a combined score | Reliability of these is usually better than their individual parts
26
Difference scores
Calculated difference between 2 scores Reliability of these is usually lower than their individual parts (information is lost: only can see change, not initial baseline)
27
Choosing a reliability test to use
Multiple administrations: test-retest reliability | One administration: homogeneous content uses coefficient alpha and heterogeneous content uses split-half coefficient
28
Factors to consider when evaluating reliability coefficients
Construct being measured Time available for testing How the scores will be used Method of estimating reliability
29
High-stake decision tests: reliability coefficient used
Greater than 0.9 or 0.95
30
General clinical use: reliability coefficient used
Greater than 0.8
31
Class tests and screening tests: reliability coefficient used
Greater than 0.7
32
How to improve reliability
Increase number of test items Use composite scores Develop better items Standardize administration
33
Standard error of measurement (SEM)
Standard deviation of test administered to the same individual an infinite number of times Useful when interpreting test scores When reliability increases, this decreases
34
How to calculate confidence intervals
Use SEM and SD
35
Relationship between reliability and confidence interval
Reliability increases, confidence interval decreases
36
Test manuals/researchers report: information included
Internal consistency Test-retest Standard error of measurement (SEM) Information on confidence intervals
37
Generalizability theory
Shows how much variance is associated with different sources of error