Variables and Their Measurement Flashcards

(65 cards)

1
Q

Variable (definition)

A

Characteristic of an individual, object, or environmental conditoin

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Types Of Variables (3)

A

1) Independent Variable
2) Dependent Variable
3) Extraneous Variable

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Independent Variable (def)

A

The variable that is intentionally manipulated by researchers to produce a change in an outcome; there are levels indicated by the number of IV’s in the study

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

IV’s are used in ___ Types of studies (2)

A
  • Interventional

- Prognostic

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

IV’s are NOT used in ___ Studies (1)

A

Descriptive; bc researchers don’t manipulate anything and there is no outcome of interest

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Purpose of IV’s in Interventional Studies (2)

A
  • Can draw CAUSE AND EFFECT relationships because manipulating an IV
  • They are commonly used as treatments/interventions
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Levels of IV’s (3)

A
  • > 1 IV = factorial design
  • 2x2 factorial design
  • 2x3 factorial design
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

2x2 factorial design?

A

Means there are 2 IV’s with 2 levels

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

2x3 factorial design?

A

Means there are 2 IV’s with 3 levels

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

IV Prognostic Studies (3)

A
  • CANNOT draw Cause and Effect relationships bc not manipulating anything!
  • IV’s are used to predict the outcome of interest
  • Not intentionally manipulated by researchers
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Dependent Variable (def)

A

The outcome of interest

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Purpose of IV’s and DV’s in Interventional Studies

A

-Manipulate IV to cause a change in the DV= CAUSAL relationship!

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

IV’s and DV’s in Prognostic Studies

A

-Assess whether the IV is associated with the DV or predicts the DV; looks at IV and tries to predict the DV

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Extraneous Variable(def)

A
  • An outside variable that influences the DV, but it’s not an IV- want to limit these!
  • Should be controlled through study design and statistical adjustment so you can say without a doubt that change was from your tx
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Discrete Variable (def)

A

Values are distinct categories

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Dichotomous Variable (def)

A
  • A type of discrete variable

- Variables who only have 2 possible values; ex. Male and Female

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

Continous Variable (def)

A

Variables who have a theoretical infinite number of values

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

Levels of Measurement (4)

A

1) Nominal
2) Ordinal
3) Interval
4) Ratio

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

Nominal Level of Measurement

A
  • Includes values that are named CATEGORIES
  • NO rank; no one category is better than another
  • Statistical analysis uses FREQUENCIES (not means)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

Ordinal Level of Measurement

A
  • Categories have a RANK order relative to one another; some form of hierarchy (ex. Likert scales)
  • Statistical Analysis uses FREQUENCIES (not means)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

Interval Level of Measurement

A
  • Assigns quantitative values to variables
  • Do NOT have a KNOWN ZERO POINT, but zero doesn’t mean there is an “absence” of the characteristic
  • Possible values extend to infinity and can be both positive or negative
  • Values may be added and subtracted, but not multiplied or divided
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

Ratio Level of Measurement

A
  • Have a RANK order, and a KNOWN ZERO POINT
  • No negative values
  • has a TRUE zero
  • Values can be added, subtracted, multiplied, and divided
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

True zero (def)

A

If there is a zero present, there is an ABSENCE of the characteristic

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
24
Q

What type of variable is MMT grades? Why?

A

Discrete and ordinal variable; not an infinite number of possibilities and there is a heirarchy

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
25
What type of variable is gender? Why?
Discrete (dichotomist) and Nominal variable; bc not an infinite number of possibilities and equal groups (no hierarchy)
26
What type of variable is Blood Pressure? Why?
Ratio Variable- bc there is a true zero (no negative numbers)
27
Reliability (def)
How consistent and free from error is the instrument; represents how much of a measurement represents error and how much is the true score
28
Measurement error (def)
Difference between observed and true score
29
Observed score =
True score-error
30
How researchers minimize measurment error?
Choose instruments which have established reliability measures
31
2 forms of Measurement Relaibility
- related to the instrument | - Related to the person TAKING the measurements
32
Types of Reliability related to the INSTRUMENT (2)
1) Reproducibility aka Test-Retest Reliability | 2) Internal Consistency
33
Reproducibility (def)
- Aka Test-Retest Reliability | - established when an instrument is used in two separate occasions with the same subjects
34
Internal Consistency (def)
Related to self-report outcome measure
35
Types of Reliability related to the RATER (2)
1) INTRA-rater reliability | 2) INTER-rater Reliability
36
Intra-rater Reliability (def)
Consistency of repeated measures performed by ONE INDIVIDUAL
37
Inter-rater Reliability (def)
Consistency of measures performed by >1 individual
38
Regression Toward the Mean
Observed scores move closer to the mean with repeated tests; affects OUTLIERS
39
Reliability Coefficient(def)
Estimates reliability based on statisical concept of variance; a Measure of variability or differences among scores within a sample
40
Reliability Coefficient (math)
- (True score variance/(true score variance + error variance)) - Ranges from 0.00-1.00
41
Reliability Coefficient that indicates POOR reliability:
42
Reliability Coefficient that indicates MODERATE reliability
.50-.75
43
Reliability Coefficient that indicates GOOD Reliability
>.75
44
Validity
Does the test measure what it's supposed to measure; goes hand-in-hand with reliability- don't want to be reliably invalid!
45
4 Types of Validity
1) Face Validity 2) Content Validity 3) Construct Validity 4) Criterion Validity
46
Face Validity (def)
- Simplest and most subjective form of validity | - The instrument appears to be the appropriate choice to measure the variable (or not)
47
Content Validity (def)
Does the measure represent all of the relevant facets of the variable of interest?; should not contain elements that capture irrelevant information
48
Construct Validity (def)
Does the measure reflect the operational definition of the concept or construct it says it represents?
49
Convergent Validity (def)
Method used to evaluate the construct validity of an instrument; assess the relationship between scores on the instrument of interest and on another instrument ; if scores from both instruments yield similar resulst, then the measure has convergent validity
50
Discriminant Validity (def)
Method used to validate construct validity of an instrument; reflects degree to which an instrument can distinguish between different constructs
51
2 Things that evaluates Construct Validity
1) Convergent Validity | 2) Discriminant Validity
52
Convergent Validity (def)
- Method used to evalute construct validity of an instrument; - Assess the relationships between scores on the instrument of interest and ON ANOTHER INSTRUMENT - If scores from both instruments yield similar results, then the measure has convergent validity
53
Discriminant Validity (def)
- Method used to validate construct validity of an instrument - Reflects the degree to which an instrument can distinguish between different constructs; ex. can a goniometer distinguish between strength and ROM?
54
Things that Evaluate Criterion Validity (2)
1) Concurrent Validity | 2) Predictive Validity
55
Criterion Validity (def)
-Degree to which the scores on an instrument are related to the scores on a reference standard instrument
56
Concurrent Validity (def)
-Method of evaluating criterion validity; administer the test of interest and reference standard test (the GOLD STANDARD) at the same time (to eliminate bias)
57
Predictive Validity (def)
- Method of evaluating criterion validity | - Degree to which the results from the test of interest can predict a future outcome
58
Sensitivity and Specificity are modes of ____ Validity
Criterion
59
A test CANNOT be ___ but not ___
valid;reliable
60
Instruments should be ____, ____, and ____ _ ____
reliable, valid, responsive to change
61
To be Responsive (to change) an instrument should Have: (2)
- Construct validity | - Severeal values on the scale to limit the floor or cieling effect
62
Standard error of Measurement (def)
Extent to which observed scores are disbursed around the "true" score
63
Large Standard of Error = Less ___; why?
Responsive; bc the "true" values are lost in the inaccuracies each time the measure is repeated
64
Ceiling effect
Happens when scores don't go high enough
65
Floor Effect
Happens when scores don't go low enough