Exam 2 Flashcards

1
Q

consistency in measurement of some (real or hypothetical) characteristic

A

reliability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

the theoretical number that it one’s perfectly accurate representation of knowledge as a score on a test

A

True Score

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

a person’s true score, plus some error

A

observed test score

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

unsystematic variability introduced into scores; random; can cause deviation from true score

A

error

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Sources of Psychological Measurement Error

A
  • test construction
  • test administration
  • test scoring
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

items are created/selected from a large population of possible items from within the domain
may affect how well one performs on a test. Someone may perform very well on some subtests and bad on others.

A

content sampling

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

factors that influence the test taker’s attention, concentration, motivation, etc.

A

test administration

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

physical appearance, departure from test standardization procedure, not placing materials in proper orientation, incorrect timing, etc.

A

examiner influences

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

test environment, test anxiety, medication effects, extended testing session - fatigue

A

physical or psychological discomfort

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

estimates of the ratio of true score variance to total variance (cannot be negative)

A

reliability coefficients

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

test-retest/stability, interrater agreement. internal consistency, alternate forms, etc.

A

types of reliability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

reliability indices should meet or exceed

A

.85 or .90

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

index of how an individual’s scores may vary over tests presumed to be parallel

A

standard error of measurement (SEm)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

as rxx increases from 0-1…

A

SEm decreases from SD to 0

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Purpose of rxx

A
  1. internal consistency
  2. alternate forms
  3. test-retest
  4. interrater agreement
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

static vs. dynamic characteristic

A

static doesn’t change very much

dynamic evolves a lot

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

degree to which evidence and theory support the interpretations of test scores for proposed uses of tests
relates to inferences or interpretations made about performance based on scores from the measure.

A

validity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q
  • content validity
  • criterion related validity
  • construct validity
A

trinitarian model

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

Evidence based on:

  • test content
  • response processes
  • internal structure
  • relations with other variables
  • consequences of testing
A

unified theory (Messick)

20
Q

examination of test content

A

content validity

21
Q
  • test-retest
  • alternate forms
  • interrater agreement
A

Methods of Reliability

22
Q

A

Cronbach’s coefficient alpha

23
Q

when an item belongs to more than one factor

A

cross loading

24
Q

examination of test content

A

content validity

25
Q
  • test-retest
  • alternate forms
  • interrater agreement
A

Methods of Reliability

26
Q

A

Cronbach’s coefficient alpha

27
Q

when an item belongs to more than one factor

A

cross loading

28
Q

z scores for 68%
95%
99%

A

1

  1. 96
  2. 58
29
Q

Which reliability estimate should be used with test-retest?

A

stability

30
Q

SEdiff formula

A

SEdiff = SD√2-rxx1-rxx2
Rxx1 = reliability of test 1
Rxx2 = reliability of test 2
answer is the number of points required to make a SIGNIFICANT DIFFERENCE

31
Q

Which reliability estimate should be used with internal consistency?

A

coefficient alpha, KR-20, split-half

32
Q

Which reliability estimate should be used with alternate forms?

A

equivalance

33
Q

Which reliability estimate should be used with test-retest?

A

stability

34
Q

compare a test in question with an already accepted standard measure

A

concurrent validity

35
Q

when correlations are superficially high because two tests are too similar

A

criterion contamination

36
Q

test in question is compared to a standard gathered at a future time; predicts performance on a standard at a future time
ex: hs GPA and ACT predicting college GPA

A

predictive validity

37
Q

experimental procedures that cause a change in construct can be used to test a test’s validity

A

theory consistent intervention effects

38
Q

two or more measures designed to measure the same construct should produce high correlations (large amount of shared variance)

A

convergent validity

39
Q

two or more measures designed to measure different constructs should produce low/near zero correlations (little shared variance)

A

divergent validity

40
Q

use scores to classify subjects into groups

A

discriminant analysis

41
Q

conceptualization

A

is there a need? who wil it be used for? content? administration? responses?

42
Q

scaling

A

how we assign “points” to responses

43
Q

try-out, analyze, revise

A

first “test” administration, analysis, revision

44
Q

Spearman brown

A

Used to estimate reliability after changing test length

45
Q

KR-20

A

Used with dichotomous