PSY Exam 2 Flashcards

(32 cards)

1
Q

What is Reliability?

A

the degree which test scores are free from errors of measurement (consistency)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is Observed Test Score?

A

true score plus error (X = T + E)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is the Classical Test Theory?

A

No instrument is perfectly reliable or consistent and all test scores contain some error

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is Measurement Error?

A

variations in measurement using a reliable instrument

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is a Reliable Test?

A

A test we can trust to measure each person in approximately the samew ay every time it is used

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What are the two types of error?

A

Random and Systematic

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is Random Error?

A

error that is random in nature that lowers the reliability of a test (will lower a test score by exactly the same amount with an infinite amount of testing)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is Systematic Error?

A

error that occurs when a source of error always changes a true score but does not lower the reliability of a test (since the test is already reliably inaccurate it affects descriptive statistics)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What is the Reliability Coefficient?

A

an index of the strength of the relationship between two sets of scores (think of spearman-brown formula)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What are the 3 categories of Reliability Coefficients?

A

Test-Retest, Alternate Forms, Internal Consistency

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is Test-Retest reliability?

A

A test is taken at two different points in time and are compared while facing the challenge of practice effects and fatigue (looking for correlation between scores)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What is Alternate Forms reliability?

A

A test developer creates two (or more) different forms of the test (order effects & parallel forms)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What is Internal Consistency reliability?

A

A measure of how related items or groups of items on a test are to each other (appropriate for homogeneous tests)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What is a Homogeneous Test?

A

a test that is measuring only one trait or characteristic

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What is a Heterogeneous Test?

A

a test that is measuring more than one trait or characteristic

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What is Scorer Reliability and Agreement?

A

the amount of consistency among scorers’ judgments (two or more individuals score the same test)

17
Q

What is interscorer agreement?

A

the amount of consistency among scorers’ judgments

18
Q

What is intrascorer reliability?

A

whether a person is consistent in the way they assigned scores from test to test

19
Q

What is Interrater agreement?

A

an index of how consistently the scorers rate or make decisions

20
Q

What is Intrarater agreement?

A

when one scorer makes consistent judgments across all tests

21
Q

What is KR-20 used for?

A

Estimating the internal consistency for tests whose questions can scored either right or wrong? (ex: true or false)

22
Q

What is coefficient alpha used for?

A

Estimating the internal consistency tests whose questions have more than two possible answers (ex: rating scales)

23
Q

What is Cohen’s Kappa?

A

an index for calculating scorer reliability / inter-rater agreement when scorers make judgments that result in nominal and ordinal data (-1.0 to 1.0)

24
Q

What is standard error of measurement?

A

an index of the amount of uncertainty or error expected in an individual’s observed test score… how much the individual’s observed score might differ from the individual’s true test score.

25
What is the standard error of measurement formula?
SEM = SD * √(1 - r) | standard error of measurement = standard deviation multiplied by the square root of 1 minus the reliability coefficient of the test
26
What is a confidence interval?
the range of scores that we feel confident will include the test taker's true score ±1 STD dev of the mean → 68.7 ±2 STD dev of the mean → 95% ±3 STD dev of the mean → 99%
27
What are the four factors that influence reliability?
test itself, test administration, test scoring, and test takers
28
What are the 6 factors related to the four sources of test error?
Test length, homogeneity of test questions, test-retest interval, test administration, scoring, and cooperation of test takers
29
What is the Generalizability Theory?
an approach for estimating reliability that is concerned with how well and under what conditions can we generalize an estimation of reliability of test scores from one test administrator to another
30
What did Dale say about G Theory?
it is a way of saying we know for reliability you have error (systematic / random) but now there are all of these sources and it will be nice to break apart these sources of error
31
What is validity?
the extent to which a test accurately measures what it is intended to measure
32