Evaluating Selection Techniques and Decisions Flashcards

(49 cards)

1
Q

The extent to which a score from a test or from an evaluation is consistent and free from error

A

Reliability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

The extent to which repeated administration of the same test will achieve similar results

A

Test-retest reliability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

The consistency of test scores across time

A

Temporal stability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

The extent to which two forms of the same test are similar

A

Alternate-forms

reliability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

A method of controlling for order effects by giving half of a sample Test A first, followed by Test B, and giving the other half of the sample Test B first, followed by Test A

A

Counterbalancing

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

The extent to which the scores on two forms of a test are similar

A

Form stability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

The extent to which responses to the same test items are consistent

A

Item stability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

The extent to which test items measure the same construct

A

Item homogeneity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

A statistic used to determine internal reliability of tests that use items with dichotomous answers (yes/no, true/false)

A
Kuder-Richardson
Formula 20 (K-R 20)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

A form of internal reliability in which the consistency of item responses is determined by comparing scores on half of the items with scores on the other half of the items

A

Split-half method

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Used to correct reliability coefficients resulting from the split-half method

A

Spearman-Brown

prophecy formula

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

A statistic used to determine internal reliability of tests that use interval or ratio scales

A

Coefficient alpha

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

The extent to which two people scoring a test agree on the test score, or the extent to which a test is scored correctly

A

Scorer reliability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

The degree to which inferences from test scores are justified by the evidence

A

Validity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

The extent to which tests or test items sample the content that they are supposed to measure

A

Content validity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

The extent to which a test score is related to some measure of job performance

A

Criterion validity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

A measure of job performance, such as attendance, productivity, or a supervisor rating

A

Criterion

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

A form of criterion validity that correlates test scores with measures of job performance for employees currently working for an organization

A

Concurrent validity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

A form of criterion validity in which test scores of applicants are compared at a later date with a measure of job performance

A

Predictive validity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

A narrow range of performance scores that makes it difficult to obtain a significant validity coefficient

A

Restricted range

21
Q

The extent to which inferences from test scores from one organization can be applied to another organization

A

Validity generalization

VG

22
Q

A form of validity generalization in which validity is inferred on the basis of a match between job components and tests previously found valid for those job components

A

Synthetic validity

23
Q

The extent to which a test actually measures the construct that it purports to measure

A

Construct validity

24
Q

A form of validity in which test scores from two contrasting groups “known” to differ on a construct are compared

A

Known-group validity

25
The extent to which a test appears to be valid
Face validity
26
Statements, such as those used in astrological forecasts, that are so general that they can be true of almost anyone
Barnum statements
27
A book containing information about the reliability and validity of various psychological tests
Mental Measurements | Yearbook (MMY)
28
A type of test taken on a computer in which the computer adapts the difficulty level of questions asked to the test taker’s success in answering previous questions
Computer-adaptive | testing (CAT)
29
A series of tables based on the selection ratio, base rate, and test validity that yield information about the percentage of future employees who will be successful if a particular test is used
Taylor-Russell tables
30
The percentage of applicants an organization hires
Selection ratio
31
Percentage of current employees who are considered successful
Base rate
32
A utility method that compares the percentage of times a selection decision was accurate with the percentage of successful employees
Proportion of correct | decisions
33
Tables that use the base rate, test validity, and applicant percentile on a test to determine the probability of future success for that applicant
Lawshe tables
34
Method of ascertaining the extent to which an organization will benefit from the use of a particular selection system
Utility formula
35
The length of time an employee has been with an organization
Tenure
36
Group differences in test scores that are unrelated to the construct being measured
Measurement bias
37
An employment practice that results in members of a protected class being negatively affected at a higher rate than members of the majority class. Adverse impact is usually determined by the four-fifths rule
Adverse impact
38
A situation in which the predicted level of job success falsely favors one group over another
Predictive bias
39
The characteristic of a test that significantly predicts a criterion for one class of people but not for another
Single-group validity
40
The characteristic of a test that significantly predicts a criterion for two groups, such as both minorities and non-minorities, but predicts significantly better for one of the two groups
Differential validity
41
A statistical procedure in which the scores from more than one criterion-valid test are weighted according to how well each test score predicts the criterion
Multiple regression
42
Selecting applicants in straight rank order of their test scores
Top-down selection
43
A method of making selection decisions in which a high score on one test can compensate for a low score on another test. For example, a high GPA might compensate for a low GRE score
Compensatory approach
44
A variation on top-down selection in which the names of the top three applicants are given to a hiring authority who can then select any of the three
Rule of three
45
The minimum test score that an applicant must achieve to be considered for hire
Passing score
46
A selection strategy in which applicants must meet or exceed the passing score on more than one selection test
Multiple-cutoff approach
47
Selection practice of administering one test at a time so that applicants must pass that test before being allowed to take the next test
Multiple-hurdle approach
48
A statistical technique based on the standard error of measurement that allows similar test scores to be grouped
Banding
49
The number of points that a test score could be off due to test unreliability
Standard error of | measurement (SEM)