Week 12: Reliability And Validity Flashcards

1
Q

What are the 3 psychometric properties

A

Reliability: reproducibility
Validity: measures what it’s supposed to measure
Responsiveness: detect change

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Reliability definition

A

The amount of error, both random and systematic, inherent in a measurement

  • AKA: reproducibility, stability, agreement, precision, association, sensitivity objectivity
  • consistency of measurement
  • a reliable measure can be expected to repeat the same score in 2 different occasions
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Measurement error

A

Eg temp, scale

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Systematic error

A

-bias
-fairly consistent that results in overestimate or underestimate
-predictable
-occur in one direction
Eg tape measure that merging at 2 instead of zero, the fire recordings will be 2 in too long
Primarily a concern of validity, not reliability (the flawed tape measure be be very consistent)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Random error

A

-errors due to chance
-unpredictable
-the primary concern of reliability
Eg: while measuring the patient: the patient moves slightly
-the therapist incorrectly reads the tape measure
-the tape measure may have some slack

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Sources or error in test performance

A

Clinician-related factors:
-random error (poor technique or interpretation)
-directional bias
Patient-related factors
-natural fluctuations
-misunderstanding of procedures or miscommunication

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Reliability types

A
  1. Rater reliability
    A) intra-rater: degree of random variation within one radar. Ie the repeated measures obtained by one person and the variation between these.
    B) inter-rater: when there is two or more examiners measuring the same thing. The variation is between the people obtaining these measures.
    2.Test-retest: repeated measurements that aren’t influence by reader
  2. Equivelance
  3. Internal consistency
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Intra-examiner reliability

A
  • scores should match when the same examiner test the same subjects in two or more occasions
  • intra-examiner reliability is the degree that the examiner agrees with himself or herself
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Inter-examiner reliability

A
  • when 2 or more examiners tests the same subjects for the same characteristic using the same measure, scores should match
  • inter-examiner reliability is the degree that their findings agree
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Quantifying inter-examiner and intra-examiner reliability

Correlation

A

There should be a high degree of correlation between scores if 2 examiners testing the same group of subjects or 1 examiner testing the same group on 2 occasions.

  • however, it is possible to have good correlation and concurrent poor agreement
    • occurs when 1 one examiner consistently scores higher or lower than other examiner
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Test-retest reliability

A

Ie if an examiner gave a subject a sheet to fill out on level of disability, then gave them the same sheet 5 minutes later then we wouldn’t expect a difference between them. If there was it would be error, but because the radar isn’t involved in this error it is a test-retest reliability .

How well did you know this?
1
Not at all
2
3
4
5
Perfectly