Chapter 5 Flashcards

(44 cards)

1
Q

Reliability (def)

A

consistency in measurement

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Reliability coefficient

A

0 to 1 statistic.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

4 types of reliability coefficients

A

1) test-retest reliability

2) alternate-forms reliability

3) split-half reliability

4) inter-scorer reliability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Measurement error (textbook def)

A

Inherent uncertainty with any measurement, even after minimizing preventable mistakes

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

2 influences that interfere with repeated measurement (in psych)

A

1) changes in object (eg. a constant flux of mood, alertness, motivation)

2) the act of measurement (i.e., carryover effects like fatigue, practice)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

“True Score”

A

not actually true to concept. True score is tied to the specific measurement instrument.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What ‘score’ measures the truth independent of measurement?

A

Construct score.

the underlying score of some construct (eg. depression)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

variance is made of what two subtypes of variance?

A

True variance (actual differences between people?) + Error variance (random variances that are irrelevant)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Define reliability in terms of variance

A

Proportion of total variance attributed to true variance

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Random vs. Systematic Error

A

Random: unpredictable, inconsistent, without pattern

Systematic: predictable, constant, can be adjusted for

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Bias (in error)

A

The degree of systematic error that influences measurement

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

How does item/content sampling contribute to error variance?

A

The specific content in some test may affect the results (eg. i hope they ask this question and not this)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What test administration effects contribute to error variance?

A

Environment: war, heat, gum, pencil, etc.

Testtaker variables: lack of sleep, emotions, drugs, etc.

Examiner-related variables: physical appearance, presence/absence

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

How does test scoring and interpretation contribute to error variance?

A

Some subjectivity in certain tests (eg. essays, creativity, etc.) can influence measurement.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

test-retest reliability coefficient is also called what?

A

Coefficient of stability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What might affect test-retest reliability estimates?

A

Experience, practice, memory, fatigue, etc. may intervene.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

alternate-forms/parallel-forms reliability estimates coefficient name

A

Coefficient of equivalence

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

Parallel vs. Alternate forms reliability

A

Parallel forms: Means and variances of test scores are equal

Alternate forms: different versions of same test, but aren’t parallel

19
Q

2 similarities between parallel/alternate and test-retest reliability

A

1) two test administrations with same group

2) test scores can be affected by factors like fatigue, practice, learning, etc.

20
Q

What additional source of error variance is present in alternate/parallel-forms reliability?

A

Item/Content sampling

21
Q

Split-half reliability

A

Correlating two pairs of scores from a single test.

one half of a test Pearson r with another half, then adjust with Spearman-Brown formula

22
Q

Odd-even reliability

A

split-half reliability by using odd vs. even numbers

23
Q

How do number of items affect reliability coefficient? What method can see how many items needed?

A

Spearman-Brown.

More items is more reliability

24
Q

What coefficient for inter-item consistency?

A

Coefficient alpha

25
Inter-scorer reliability What coefficient?
Degree of consistency between 2 or more scorers. Coefficient of inter-scorer reliability
26
DSM-5 Inter-rater reliability
Kappa = 0.44 (fair level moderately greater than chance)
27
Transient error
Error due to testtaker's feelings, moods, or mental state over time
28
Homogeneity vs. Heterogeneity of test items
Homogenous: Functionally uniform items. Measures one factor (eg. one ability/trait). High internal consistency should happen Heterogenous: Not just one factor measured in the test.
29
Does high internal consistency mean homogeneity of items?
Not necessarily. More items will lead to high internal consistency coefficients as long as they're positively correlated
30
Dynamic vs. static characteristics
Dynamic: Presumed to be relatively situational and changing Static: presumed to be relatively unchanging
31
Restriction/Inflation of range
When some subgroup inflates or restricts the correlational analysis??
32
Power Test
Enough time to attempt all items, but so difficult that nobody gets perfect score
33
Speed test
Same level of difficulty in items and testtakers should complete everything correctly if unlimited time. But only some will be able to complete the whole test
34
What's differences in assumptions between CTT and IRT? (not specific, but ya..)
CTT assumptions are weak and easily met. IRT are rigorous.
35
Domain Sampling Theory
Reliability is based on how well a score assesses the domain of where a sample is drawn.
36
What is universe score in generalizability theory?
The true score (given same conditions, the same score will be obtained)
37
Generalizability Study Coefficient of generalizability
how generalizable scores from a particular test are if administered in different situations.
38
Decision study
Usefulness of test scores in helping user make decisions. Follows generalizability study
39
Another way to say Item response theory
Latent-trait theory
40
Within CTT, what is the weight assigned to each item on a test?
Equal weight. IRT is differentital weight.
41
Dichotomous test items
can only answer with one of two responses
42
Polytomous test items
3 or more alternative responses
43
Rasch Model
a type of IRT model with underlying distribution assumption
44
Which measure is used to compare differences between scores?
Standard error of the difference