Chapter 12 & 13 - Measurement of Variables Flashcards

1
Q

Four levels of measurement

A

Nominal
Ordinal
Interval
Ratio

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Categorical levels of measurement

A

Nominal
Ordinal

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Continuous levels of measurement

A

Interval
Ratio

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Nominal level of measurement

A

Identification; classification

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Ordinal level of measurement

A

Ranking of categories, but not equidistant (at equal distances)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Interval level of measurement

A
  • Equidistant (at equal distances) ranking, but zero point not fixed
  • Likert-type scales: Ordinal or interval?
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Ratio level of measurement

A

Possesses a unique origin (zero point)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Rating Scale Formats

A

– Dichotomous/binary scale
– Category scale
– Semantic differential scale
– Numerical scale
– Itemized rating scale
– Likert-type scale
– Fixed or constant sum scale
– Stapel scale
– Consensus scale
– Graphic rating scale
– Paired comparison scale
– Forced choice scale
– Comparative scale

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Dichotomous/binary scale rating scale example

A

Yes vs. No

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Category scale rating scale example

A

English; French; Other

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Semantic differential scale rating scale example

A

Good —– Bad; Emotionally stable —– Neurotic

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Numerical scale rating scale example

A

Responsive 1 2 3 4 5 6 7 Unresponsive

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Itemized rating scale rating scale example

A

Can be balanced or unbalanced; forced or unforced

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Likert-type scale rating scale example

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Fixed or constant sum scale rating scale example

A

distributing 100 points across several items (need to add
up to 100)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Stapel scale rating scale example

A

-3 -2 -1 Interpersonal skills +1 +2 +3

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

Consensus scale rating scale example

A

Developed by consensus by a panel of judges

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

Graphic rating scale rating scale example

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

Paired comparison scale rating scale example

A

Respondents asked to choose between two objects at a time
(among a small number of objects)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

Forced choice scale rating scale example

A

Ranking objects among the provided alternatives

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

Comparative scale rating scale example

A

Provides a point of reference to assess attitudes toward a
particular object/event/situation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

Response-scale format considerations

A

– Measurement scale (nominal, ordinal, interval, ratio)
– Number of scale points/categories
* Need to be mutually exclusive and collectively exhaustive
– Balanced or unbalanced scales
* Equal # of favourable & unfavourable categories?
– Forced or non-forced choice
* “Neutral” category
* Odd or even number of categories
– Category labels for scales (anchors)
* Verbal; numerical; unlabeled (e.g., graphic)
– Number of items in a scale

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

Operationalization of variables

A

Breaking an abstract construct down to its measureable or
tangible components
* Can tap into a construct by looking at the behavioural dimensions, facets, and properties denoted by the construct and translating them into observable/measurable elements

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
24
Q

Delineating the antecedents, consequences, or correlates of the construct is not _______________

A

operationalization

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
25
Steps in the Operationalization of variables
* Clear definition of the construct (and possibly its dimensions) * Develop a pool of items (indicators or elements) representing more concrete manifestations or operalizations of the construct * Choose a response format (e.g., Likert-type scale) * Collect data from a (representative) sample * Conduct item analyses and select items for the scale(s) * Test the reliability and validity of the scale(s)
26
Typical process in Developing Scales
– Define the concept to be measured – Identify components/elements of the concept – Specify sample of observable, measurable items representing the components/elements of the concept – Select an appropriate response format to measure the items – Combine the items into a composite (summated) scale – Pretest the scale to assess respondent understanding – Assess scale reliability and validity – Revise the scale (if needed)
27
considerations for evaluating existing scales
– Title, author(s), publisher (if applicable) – Language(s); equivalence of translated forms – Construct(s) (allegedly) measured – Characteristics of development/normative sample(s) – Costs/permissions required – User qualifications – Format; administration method; scoring; length/time – Psychometrics: Evidence of reliability, validity, fairness – Independent reviews or peer-reviewed research
28
____________ Consist of a number of closely related items (questions or statements) whose responses are combined into a composite score to measure a construct
Multi-Item (Summated) Scales
29
Recommendations for Multi-Item (Summated) Scales
Items should be closely related, represent only a single construct, and represent the construct completely
30
Assessing Measurement Scales
Reliability Validity (unitary)
31
Assessing Measurement Scales - Reliability
* Test-retest reliability * Parallel-form reliability * Inter-rater reliability * Internal consistency reliability * Split-half reliability
32
Assessing Measurement Scales - Validity (unitary)
* Content validity * Construct validity * Convergent validity * Discriminant validity * Criterion-related validity * Concurrent validity * Predictive validity
33
Reliability
the stability and consistency of scores generated by a scale
34
Stability of scores (derived from scales)
– Test-retest reliability (stability over time) – Parallel-form reliability (stability across forms) – Administer them to the same subjects – Inter-rater reliability (stability across raters)
35
Inter-item (or internal) consistency reliability
* Consistency of answers to all the items in a measure – If these items independently measure the same concept, they will be highly correlated * Coefficient α or Cronbach’s α – A different “alpha” than the one associated with Type I error * Rule of thumb: Use α >.70 as “acceptable” in our field
36
Split-half reliability
* Correlations between two halves of an instrument * Typically not as useful as Cronbach’s α
37
Validity
whether an instrument measures what it sets out to measure
38
Construct validity
Degree of correspondence between a construct and its operational definition (measure or manipulation)
39
Forms of evidence in validity
* Content-based (content validity) * Convergent and discriminant validity * Criterion-related evidence (concurrent and predictive)
40
Content validity
Evidence that the content of a test corresponds to the content of the construct it was designed to measure * Usually relies on opinions of subject matter experts
41
Face validity
Does the scale appear to measure the construct?
42
Convergent validity
– Identify another scale that measures the same construct as the one being validated – Obtain scores on both scales and compute the correlation between them (should be high)
43
Discriminant validity
– Identify a scale that measures a different construct – Specify how the two scales are expected to differ – Obtain scores on both scales and compute the correlation between them (should be low)
44
Convergent and discriminant validity
Professional credentialing examinations are designed to assess the knowledge required for competent professional practice in a given discipline.
45
Measurement
The assignment of numbers or other symbols to characteristics (or attributes) of objects according to a prespecified set of rules.
46
Operationalizing
Reduction of abstract concepts to render them measurable in a tangible way.
47
Some considerations for evaluating existing scales
– Title, author(s), publisher (if applicable) – Language(s); equivalence of translated forms – Construct(s) (allegedly) measured – Characteristics of development/normative sample(s) – Costs/permissions required – User qualifications – Format; administration method; scoring; length/time – Psychometrics: Evidence of reliability, validity, fairness – Independent reviews or peer-reviewed research
48
Multi-Item (Summated) Scales
Consist of a number of closely related items (questions or statements) whose responses are combined into a composite score to measure a construct – Scale / Index / Summated rating scale / Multi-item scale
49
Scale
A tool or mechanism by which individuals, events, or objects are distinguished on the variables of interest in some meaningful way.
50
Likert scale
An interval scale that specifically uses the five anchors of Strongly Disagree, Disagree, Neither Disagree nor Agree, Agree, and Strongly Agree.
51
Nominal scale
A scale that categorizes individuals or objects into mutually exclusive and collectively exhaustive groups, and offers basic, categorical information on the variable of interest.
52
Ordinal scale
A scale that not only categorizes the qualitative differences in the variable of interest, but also allows for the rank‐ordering of these categories in a meaningful way.
53
Interval Scale
A multipoint scale that taps the differences, the order, and the equality of the magnitude of the differences in the responses
54
Ratio scale
A scale that has an absolute zero origin, and hence indicates not only the magnitude, but also the proportion, of the differences.
55
Rating scale
Scale with several response categories that evaluate an object on a scale.
56
Ranking Scale
Scale used to tap preferences between two or among more objects or items
57
Dichotomous scale
Scale used to elicit a Yes/No response, or an answer to two different aspects of a concept.
58
Category scale
A scale that uses multiple items to seek a single response.
59
Semantic differential scale
Usually a seven‐point scale with bipolar attributes indicated at its extremes.
60
Numerical scale
A scale with bipolar attributes with five points or seven points indicated on the scale.
61
Unbalanced rating scale
An even‐numbered scale that has no neutral point.
62
Faces scale
A particular representation of the graphic scale, depicting faces with expressions that range from smiling to sad.
63
Consensus scale
A scale developed through consensus or the unanimous agreement of a panel of judges as to the items that measure a concept.
64
Constant sum rating scale
A scale where the respondents distribute a fixed number of points across several items.
65
Paired comparisons
Respondents choose between two objects at a time, with the process repeated with a small number of objects.
66
Forced choice
Elicits the ranking of objects relative to one another.
67
Comparative scale
A scale that provides a benchmark or point of reference to assess attitudes, opinions, and the like.
68
Reliability
Attests to the consistency and stability of the measuring instrument.
69
Validity
Evidence that the instrument, technique, or process used to measure a concept does indeed measure the intended concept.
70
Goodness of measures
Attests to the reliability and validity of measures.
71
Content validity
Establishes the representative sampling of a whole set of items that measures a concept, and reflects how well the dimensions and elements thereof are delineated.
72
Face validity
An aspect of validity examining whether the item on the scale, on the face of it, reads as if it indeed measures what it is supposed to measure.
73
Criterion-related validity
That which is established when the measure differentiates individuals on a criterion that it is expected to predict.
74
Concurrent validity
Relates to criterion‐related validity, which is established at the same time the test is administered.
75
Predictive validity
The ability of the measure to differentiate among individuals as to a criterion predicted for the future.
76
Construct validity
Testifies to how well the results obtained from the use of the measure fit the theories around which the test was designed.
77
Convergent validity
That which is established when the scores obtained by two different instruments measuring the same concept, or by measuring the concept by two different methods, are highly correlated.
78
Discriminant validity
That which is established when two variables are theorized to be uncorrelated, and the scores obtained by measuring them are indeed empirically found to be so.
79
Test–retest reliability
A way of establishing the stability of the measuring instrument by correlating the scores obtained through its administration to the same set of respondents at two different points in time.
80
Parallel‐form reliability
That form of reliability which is established when responses to two comparable sets of measures tapping the same construct are highly correlated.
81
Internal consistency
Homogeneity of the items in the measure that tap a construct.
82
Inter item consistency reliability
A test of the consistency of responses to all the items in a measure to establish that they hang together as a set.
83
Split‐half reliability
The correlation coefficient between one half of the items measuring a concept and the other half.
84
Reflective scale
Each item in a reflective scale is assumed to share a common basis (the underlying construct of interest).
85
Formative scale
Used when a construct is viewed as an explanatory combination of its indicators.
86