Validity and Reliability Flashcards

1
Q

What is a theory?

A

A body of interrelated principles that present a systematic of phenomena

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is a hypothesis?

A

A statement of the expected relationship between variables

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What 4 things are generally referred to as variables?

A

1) tests
2) diagnoses
3) treatments
4) effects of interest in a research question

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What are the 2 reference standards?

A

1) norm-referenced

2) criterion-referenced

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What are norm-referenced standards?

A

Standard that judge’s an individual’s performance off of the group’s performance
*Growth curves for children are examples of clinical measurements that are norm-referenced

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What are criterion-referenced standards?

A

Standard that an individual’s performance off of an absolute standard
*Discharge criteria such as “transfers independently” or “ambulates 300 ft” are examples of clinical situations that are criterion-referenced

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is measurement reliability?

A

The amount of variability in a measure (consistency)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Measurement reliability have different forms primarily classified by what two things?

A

instrument and rater (person or people taking the measurements)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

3 Measurement Reliability for Instruments

A

1) Internal Consistency
2) Parallel Forms
3) Spilt-Half

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What is internal consistency reliability?

A

Refers to the degree to which tests or procedures measure the same concept or construct
The instrument should demonstrate internal consistency for each construct
*Health-related quality of life questionnaires for example

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is parallel forms reliability?

A

Used to assess consistency of the results of two tests constructed in the same way from the same content domain

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What is split-half reliability?

A

The reliability of an instrument established by testing two versions of the same tool that are combined into one survey administered at one time
In other words administer the entire instrument to a sample of people and calculate the total score for each randomly divided half and determine the degree to which scores agree

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

2 Measurement Reliability for Raters

A

1) Inter-rater reliability

2) Intra-rater reliability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What is inter-rater reliability?

A

Refers to the consistency of measurement values between different individual raters
*If physical therapists from the same clinic take turns collecting data for a study it is considered inter-rater reliability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What is intra-rater reliability?

A

Refers to the consistency of measurement values of one individual
*A single PT is responsible for measuring joint ROM following an experimental stretching technique should be able to obtain nearly the same score for the same position each time the measurement is taken

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What is Measurement Validity?

A

The ability of a measure to capture what it is intended to capture
For example: a goniometer that measures joint position in degrees is a valid instrument for ROM, whereas a thermometer is not

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

How is validity strengthened?

A

by the decisions made in a design or methodology

18
Q

How is validity threatened?

A

by decisions made in design or methodology

19
Q

4 Forms of Measurement Validity

A

1) Face validity
2) Content Validity
3) Construct Validity
4) Criterion validity

20
Q

Describe Face Validity

A

It is the simplest and weakest form of measurement validity. It is typically tested with a yes or no question such as “Does this instrument appear to be the appropriate choice to measure this variable?”

21
Q

Describe Content Validity

A

An instrument has content validity if it represents all of the relevant facets of the variable it intends to measure
*Fitness variables for example

22
Q

Why is content validity challenging to determine?

A

Because there is no external standard or statistical criterion against which an instrument can be judged

23
Q

What is construct validity?

A

This type of instrument validity is determined based on the degree to which the measure reflects the definition of the concept or construct that it represents
For example: patient satisfaction surveys

24
Q

Why is construct validity difficult to achieve?

A

Because definitions evolve over time through repeated testing that changes people’s understanding of the theoretical basis underlying the concept or construct of interest

25
Q

What is Criterion Validity?

A

This type of instrument validity reflects the degree to which its scores are related to scores obtained with a reference standard instrument
For example: clinician examination techniques will have high criterion validity if they produce the same results of the MRI

26
Q

What are the two methods of evaluating criterion validity?

A

1) Concurrent Validity

2) Predictive Validity

27
Q

What is concurrent validity?

A

a method of evaluating criterion validity that involves administering the test of interest and reference standard test at the same time.

28
Q

What is predictive validity?

A

a method of evaluating criterion validity that reflects the degree to which the results form the test of interest can predict a future outcome, preferably measured by a reference standard

29
Q

What is responsiveness?

A

The ability of a measure to detect change in the phenomenon of interest

30
Q

What 3 things does responsiveness to change depend on?

A

1) the fit between the instrument and the operational definition of the phenomenon interest
2) the number of values on the instrument scale
3) the standard error of measurement associated with the instrument

31
Q

What is the first requirement for a responsive instrument?

A

construct validity: the measure should match the operational definition of the phenomenon of interest

32
Q

What is the second requirement for a responsive instrument?

A

The more values on a scale the more responsiveness to change

33
Q

What is the third requirement for a responsive instrument?

A

Standard error of measurement (SEM) is extent to which observed scores are disbursed around “true score”

34
Q

An instrument with a large standard error of measurement (SEM) will be ____ responsive to change.

A

Less

35
Q

When do floor and ceiling effects occur?

A

When the scale of the measure foes not register a further decrease or increase in scores for the lowest or highest scoring individuals

36
Q

Describe floor effect

A

There is no change is seen in scores even with lower performance
For Example: Pain is increasing but the therapist cannot see it

37
Q

Describe Ceiling effect

A

There is no change is seen in scores even with better performance
For Example: Patient feels better but there is no increase in the scores

38
Q

What is interpreting change mean?

A

The ability of an instrument to detect meaningful change is an essential criterion for its selection

39
Q

Internal vs. External Validity

A

Internal validity occurs when a researcher controls all extraneous variables and the only variable influencing the results of a study is the one being manipulated by the researcher. (MOST IMPORTANT)

External Validity refers to the extent to which the results of a study can be generalized or extended to others outside the experimental situation

40
Q

When does the internal validity of a study decrease?

A

If there is no control group, not enough subjects, you use faulty instruments, etc.