CHAPTER 6: EVALUATING SELECTION TECHNIQUES AND DESCISIONS Flashcards

1
Q

The extent to which a score from a test or from an evaluation is consistent and free from error.

A

Reliability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

The extent to which repeated administration of the same test will achieve similar results.

A

Test-retest reliability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

The consistency of test scores across time.

A

Temporal stability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

The extent to which two forms of the same test are similar.

A

Alternate-forms reliability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

A method of controlling for order effects by giving half of a sample Test A first, followed by Test B, and giving the other half of the sample Test B first, followed by Test A.

A

Counterbalancing

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

The extent to which the scores on two forms of a test are similar.

A

Form stability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

The extent to which responses to the same test items are consistent.

A

Item stability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

The extent to which test items measure the same construct.

A

Item homogeneity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

A statistic used to determine internal reliability of tests that use items with dichotomous answers (yes/no, true/false).

A

Kuder-Richardson Formula 20 (K-R 20)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

A form of internal reliability in which the consistency of item responses is determined by comparing scores on half of the items with scores on the other half of the items.

A

Split-half method

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Used to correct reliability coefficients resulting from the split-half method.

A

Spearman-Brown prophecy formula

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

A statistic used to determine internal reliability of tests that use interval or ratio scales.

A

Coefficient alpha

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

The extent to which two people scoring a test agree on the test score, or the extent to which a test is scored correctly.

A

Scorer reliability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

The degree to which inferences from test scores are justified by the evidence.

A

Validity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

The extent to which tests or test items sample the content that they are supposed to measure.

A

content validity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

The extent to which a test score is related to some measure of job performance.

A

Criterion validity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

A measure of job performance, such as attendance, productivity, or a supervisor rating.

A

Criterion

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

A form of criterion validity that correlates test scores with measures of job performance for employees currently working for an organization

A

Concurrent validity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

A form of criterion validity in which test scores of applicants are compared at a later date with a measure of job performance.

A

Predictive validity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

A narrow range of performance scores that makes it difficult to obtain a significant validity coefficient.

A

Restricted range

21
Q

The extent to which inferences from test scores from one organization can be applied to another organization.

A

Validity generalization (VG)

22
Q

A form of validity generalization in which validity is inferred on the basis of a match between job components and tests previously found valid for those job components.

A

Synthetic validity

23
Q

The extent to which a test actually measures the construct that it purports to measure.

A

Construct validity

24
Q

A form of validity in which test scores from two contrasting groups “known” to differ on a construct are compared.

A

Known-group validity

25
The extent to which a test appears to be valid.
Face validity
26
Statements, such as those used in astrological forecasts, that are so general that they can be true of almost anyone.
Barnum statements
27
A book containing information about the reliability and validity of various psychological tests.
Mental Measurements Yearbook (MMY)
28
A type of test taken on a computer in which the computer adapts the difficulty level of questions asked to the test taker’s success in answering previous questions.
Computer-adaptive testing (CAT)
29
A series of tables based on the selection ratio, base rate, and test validity that yield information about the percentage of future employees who will be successful if a particular test is used.
Taylor-Russell tables
30
The percentage of applicants an organization hires.
Selection ratio
31
Percentage of current employees who are considered successful.
Base rate
32
A utility method that compares the percentage of times a selection decision was accurate with the percentage of successful employees.
Proportion of correct decisions
33
Tables that use thebase rate, test validity, andapplicant percentileona test todeterminetheprobability of future success for that applicant.
Lawshe tables
34
Method of as certaining the extent to which an organization will benefit from the use of a particular selection system.
Utility formula
35
The length of time an employee has been with an organization.
Tenure
36
Group differences in test scores that are unrelated to the construct being measured.
Measurement bias
37
An employment practice that results in members of a protected class being negatively affected at a higher rate than members of the majority class. Adverse impact is usually determined by the fourfifths rule.
Adverse impact
38
A situation in which the predicted level of job success falsely favors one group over another.
Predictive bias
39
The characteristic of a test that significantly predicts a criterion for one class of people but not for another.
Single-group validity
40
The characteristic of a test that significantly predicts a criterion for two groups, such as both minorities and nonminorities, but predicts significantly better for one of the two groups.
Differential validity
41
A statistical procedure in which the scores from more than one criterion-valid test are weighted according to how well each test score predicts the criterion.
Multiple regression
42
Selecting applicants in straight rank order of their test scores.
Top-down selection
43
A method of making selection decisions in which a high score on one test can compensate for a low score on another test. For example, a high GPA might compensate for a low GRE score.
Compensatory approach
44
A variation on top-down selection in which the names of the top three applicants are given to a hiring authority who can then select any of the three.
Rule of three
45
The minimum test score that an applicant must achieve to be considered for hire.
Passing score
46
A selection strategy in which applicants must meet or exceed the passing score on more than one selection test.
Multiple-cutoff approach
47
Selection practice of administering one test at a time so that applicants must pass that test before being allowed to take the next test.
Multiple-hurdle approach
48
A statistical technique based on the standard error of measurement that allows similar test scores to be grouped.
Banding
49
The number of points that a test score could be off due to test unreliability.
Standard error of measurement (SEM