Prelim 2 prep Flashcards

1
Q

What are the differences between True Score Theory, Generalizability Theory, and Item Response Theory?

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is the standard error of measurement?

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What are confidence intervals and what do they tell us?

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

When confidence intervals increase in terms of percentage (i.e., 90% vs 95%), what does that do
to the range of scores it comprises?

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What does it mean if a test is valid?

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What are the three main categories of validity?

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is: content validity, criterion-related validity, construct validity, ecological validity, external
validity, face validity?

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is the content validity ratio and how is it used to determine content validity of test items?

A

If more than half the panelists of experts say an item is essential, has content validity, CVR 0 is half, negative if fewer than half

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Name the three characteristics of a criterion

A

Relevant, valid, uncontaminated

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What does it mean for a criterion to be uncontaminated?

A

Independent- independent group of raters decides who is good and who isn’t, then correlate that with test scores

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Define concurrent validity and predictive validity

A

Concurrent- degree to which a test score is related to some criterion measure obtained at the same time
Predictive- degree to which a test score predicts a criterion measure

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What are false negatives, false positives, specificity, and sensitivity?

A

False negative- test predicts someone doesn’t possess a trait and they do
False positive- test says someone has a trait and they don’t
Specificity- perfect wouldn’t mistakenly identify as someone having a trait when they don’t
Sensitivity- perfect identify all people who have the trait

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What is incremental validity and what would be proof of its existence?

A

Extent to which adding a second or third predictor gives more information about a criterion
proof??

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What is construct validity?

A

Extent to which a test measures a construct we are examining

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Name a describe the several ways in which you can find evidence for construct validity.

A
  • Homogenous
  • Evidence changes w age
  • Test scores change w experience
  • Distinct groups score differently
  • Convergent evidence between two tests measuring the same construct
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What is the difference between convergent and concurrent validity?

A
17
Q

What is a factor analysis and how does an exploratory factor analysis differ from a confirmatory
one?

A

?
Exploratory- estimating or extracting factors, deciding how many to retain, rotating to an interpretable orientation ????
Confirmatory- degree to which a hypothetical model fits the data

18
Q

Name and define the different types of rating error that can occur

A

Leniency error- arises from tendency on part of rater to be lenient
Severity- opposite
Central tendency- rater doesn’t use extreme ends of scale
Halo effect- seeing people well no matter what

19
Q

What is test utility?

A

Usefulness or practical value of testing to improve efficiency, use in a particular situation helps us make better decisions

20
Q

What are some of the costs of administering a test, and what are some costs of NOT
administering one?

A

Administering:
- buying
- supply of blank test protocols
- computer program to score the test
- paying to score the test
- hiring people to administer the test
- costs of doing business

Not administering:
- loss of confidence as an ultimate cost of the company???
- missing a child abuser
- failing to diagnose when someone underreports on an interview???

21
Q

Keep in mind the real-life example I discussed about how to think about the cost of testing when
doing evaluations.

A

????

22
Q

What are the Taylor Russel tables used for and what three variables are considered when using
them to decide if giving a test is “worth it.”

A
  • COME BACK
23
Q

Be able to name a few other tables (i.e., Naylor-Shine) and have a basic sense of how they work
(they could be multiple-choice option for instance)

A

COME BACK

24
Q

Name some different ways cut scores are determined.

A

COME BACK

25
Q

What’s the difference between a fixed and relative cut score?

A

Relative- actual score you need to meet a criteria changes
Fixed- always the same

26
Q

What is pilot work and why is it used?

A

Preliminary research surrounding creation of prototype of test, experiment with test items

27
Q

Name some ways scales are graded?

A

Age based
Grade based
Unidimensional v multidimensional??
Categorical v dimensional

28
Q

What are some scaling methods – and remember, they can overlap – so a categorical scale can
be graded “summatively,” etc.

A

Rating
Summative
Paired comparisons
Sorting tasks
Categorical scaling
Guttman scale
????

29
Q

What is the empirical vs analytical way of writing test items?

A

Analytical- write test questions you think will measure the qualities you want to measure
Empirical- find people with a problem, ask different types of questions, see how they respond

30
Q

Why would we want to find seemingly arbitrary items for use of distinguishing one group for
another (in other words, non-face-valid ones)

A

?

31
Q

Name some different ways items can be formatted.

A

Selected response
Constructed response
Computerized adaptive testing?

32
Q

What is computerized adaptive testing, and how does item branching work?

A

?
Adds or deletes a branch depending on performance (meaning what)?

33
Q

Name and describe a few different ways in which items can be scored

A

Cumulative
Class/category scoring
Ipsative scoring- comparing one score on one scale to another scale within the same test ?

34
Q

What are the following: item-difficulty index, item endorsement index, item reliability index,
item discrimination index

A

?

35
Q

How are item characteristic curves useful?

A

?

36
Q

Extra paper

A

almost done