CHAPTER 6: EVALUATING SELECTION TECHNIQUES AND DESCISIONS Flashcards
The extent to which a score from a test or from an evaluation is consistent and free from error.
Reliability
The extent to which repeated administration of the same test will achieve similar results.
Test-retest reliability
The consistency of test scores across time.
Temporal stability
The extent to which two forms of the same test are similar.
Alternate-forms reliability
A method of controlling for order effects by giving half of a sample Test A first, followed by Test B, and giving the other half of the sample Test B first, followed by Test A.
Counterbalancing
The extent to which the scores on two forms of a test are similar.
Form stability
The extent to which responses to the same test items are consistent.
Item stability
The extent to which test items measure the same construct.
Item homogeneity
A statistic used to determine internal reliability of tests that use items with dichotomous answers (yes/no, true/false).
Kuder-Richardson Formula 20 (K-R 20)
A form of internal reliability in which the consistency of item responses is determined by comparing scores on half of the items with scores on the other half of the items.
Split-half method
Used to correct reliability coefficients resulting from the split-half method.
Spearman-Brown prophecy formula
A statistic used to determine internal reliability of tests that use interval or ratio scales.
Coefficient alpha
The extent to which two people scoring a test agree on the test score, or the extent to which a test is scored correctly.
Scorer reliability
The degree to which inferences from test scores are justified by the evidence.
Validity
The extent to which tests or test items sample the content that they are supposed to measure.
content validity
The extent to which a test score is related to some measure of job performance.
Criterion validity
A measure of job performance, such as attendance, productivity, or a supervisor rating.
Criterion
A form of criterion validity that correlates test scores with measures of job performance for employees currently working for an organization
Concurrent validity
A form of criterion validity in which test scores of applicants are compared at a later date with a measure of job performance.
Predictive validity
A narrow range of performance scores that makes it difficult to obtain a significant validity coefficient.
Restricted range
The extent to which inferences from test scores from one organization can be applied to another organization.
Validity generalization (VG)
A form of validity generalization in which validity is inferred on the basis of a match between job components and tests previously found valid for those job components.
Synthetic validity
The extent to which a test actually measures the construct that it purports to measure.
Construct validity
A form of validity in which test scores from two contrasting groups “known” to differ on a construct are compared.
Known-group validity