Prelim 2 prep Flashcards

(36 cards)

1
Q

What are the differences between True Score Theory, Generalizability Theory, and Item Response Theory?

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is the standard error of measurement?

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What are confidence intervals and what do they tell us?

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

When confidence intervals increase in terms of percentage (i.e., 90% vs 95%), what does that do
to the range of scores it comprises?

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What does it mean if a test is valid?

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What are the three main categories of validity?

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is: content validity, criterion-related validity, construct validity, ecological validity, external
validity, face validity?

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is the content validity ratio and how is it used to determine content validity of test items?

A

If more than half the panelists of experts say an item is essential, has content validity, CVR 0 is half, negative if fewer than half

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Name the three characteristics of a criterion

A

Relevant, valid, uncontaminated

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What does it mean for a criterion to be uncontaminated?

A

Independent- independent group of raters decides who is good and who isn’t, then correlate that with test scores

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Define concurrent validity and predictive validity

A

Concurrent- degree to which a test score is related to some criterion measure obtained at the same time
Predictive- degree to which a test score predicts a criterion measure

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What are false negatives, false positives, specificity, and sensitivity?

A

False negative- test predicts someone doesn’t possess a trait and they do
False positive- test says someone has a trait and they don’t
Specificity- perfect wouldn’t mistakenly identify as someone having a trait when they don’t
Sensitivity- perfect identify all people who have the trait

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What is incremental validity and what would be proof of its existence?

A

Extent to which adding a second or third predictor gives more information about a criterion
proof??

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What is construct validity?

A

Extent to which a test measures a construct we are examining

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Name a describe the several ways in which you can find evidence for construct validity.

A
  • Homogenous
  • Evidence changes w age
  • Test scores change w experience
  • Distinct groups score differently
  • Convergent evidence between two tests measuring the same construct
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What is the difference between convergent and concurrent validity?

17
Q

What is a factor analysis and how does an exploratory factor analysis differ from a confirmatory
one?

A

?
Exploratory- estimating or extracting factors, deciding how many to retain, rotating to an interpretable orientation ????
Confirmatory- degree to which a hypothetical model fits the data

18
Q

Name and define the different types of rating error that can occur

A

Leniency error- arises from tendency on part of rater to be lenient
Severity- opposite
Central tendency- rater doesn’t use extreme ends of scale
Halo effect- seeing people well no matter what

19
Q

What is test utility?

A

Usefulness or practical value of testing to improve efficiency, use in a particular situation helps us make better decisions

20
Q

What are some of the costs of administering a test, and what are some costs of NOT
administering one?

A

Administering:
- buying
- supply of blank test protocols
- computer program to score the test
- paying to score the test
- hiring people to administer the test
- costs of doing business

Not administering:
- loss of confidence as an ultimate cost of the company???
- missing a child abuser
- failing to diagnose when someone underreports on an interview???

21
Q

Keep in mind the real-life example I discussed about how to think about the cost of testing when
doing evaluations.

22
Q

What are the Taylor Russel tables used for and what three variables are considered when using
them to decide if giving a test is “worth it.”

23
Q

Be able to name a few other tables (i.e., Naylor-Shine) and have a basic sense of how they work
(they could be multiple-choice option for instance)

24
Q

Name some different ways cut scores are determined.

25
What’s the difference between a fixed and relative cut score?
Relative- actual score you need to meet a criteria changes Fixed- always the same
26
What is pilot work and why is it used?
Preliminary research surrounding creation of prototype of test, experiment with test items
27
Name some ways scales are graded?
Age based Grade based Unidimensional v multidimensional?? Categorical v dimensional
28
What are some scaling methods – and remember, they can overlap – so a categorical scale can be graded “summatively,” etc.
Rating Summative Paired comparisons Sorting tasks Categorical scaling Guttman scale ????
29
What is the empirical vs analytical way of writing test items?
Analytical- write test questions you think will measure the qualities you want to measure Empirical- find people with a problem, ask different types of questions, see how they respond
30
Why would we want to find seemingly arbitrary items for use of distinguishing one group for another (in other words, non-face-valid ones)
?
31
Name some different ways items can be formatted.
Selected response Constructed response Computerized adaptive testing?
32
What is computerized adaptive testing, and how does item branching work?
? Adds or deletes a branch depending on performance (meaning what)?
33
Name and describe a few different ways in which items can be scored
Cumulative Class/category scoring Ipsative scoring- comparing one score on one scale to another scale within the same test ?
34
What are the following: item-difficulty index, item endorsement index, item reliability index, item discrimination index
?
35
How are item characteristic curves useful?
?
36
Extra paper
almost done