Test Construction Flashcards
(40 cards)
Classical test theory
- measurement that is used to develop and evaluate tests
- framework
- assumes that obtain test scores (X) are due to the combination of true variability (T) and measurement error (E)
X=T+E
True score variability
- result of actual differences among examinees in regards to whatever the test is measuring
- assumed to be consistent
Measurement error
- due to random factors that affect test performance of examinees in unpredictable ways
Example: distractions, ambiguously worded test items, and examinee fatigue
Test reliability
- The extent to which a test provides consistent information
- there are several methods for evaluating this and each is appropriate for different circumstances
+ Most methods provide a reliability coefficient
Reliability coefficient
- type of correlation coefficient
- ranges from 0 to 1.0
- designated with the letter r with the subscript of two of the same letters or numbers
- always interpreted as the the direct amount of variability obtained in test scores that’s due to true score variability
- .70 or higher is considered the minimally acceptable level, but 90 is usually required for higher stakes tests used to select employees, assign diagnoses, or other important decisions about individuals
The acceptable level of reliability
- depends on the type of test and its purpose
Standardized cognitive ability tests versus personality test
- cognitive ability test have a higher reliability coefficient
Standardized cognitive ability tests versus personality test
- cognitive ability test have a higher reliability coefficient
Test retest reliability
- provides information on the consistency of scores over time
- involves administering the test to a sample of examinees and readministering the test to the same examinees at a later time than correlating the two sets of scores
- useful for tests that are designed to measure a characteristic that is stable over time
Alternative forms reliability
,- provides information about the consistency of scores over different forms of the test and when the second form is administered at a later time, the consistency of scores over time
- involves administering one forum to a sample of examinees administering the other forms to the same sample of examinees and correlating the sets of scores
- important whenever a test has more than one form
Internal consistency reliability
- provides information on the consistency of scores over different test items
- useful for tests that are designed to measure a single content, domain or aspect of behavior
,- not useful for speed tests because it overestimates their reliability
Speed test
- test retest reliability and alternative forms reliability are appropriate
Coefficient alpha
-aka cronbach’s alpha
- involves administering the test to a sample of examinees and calculating the average inter-item consistency
Coefficient alpha
-aka cronbach’s alpha
- involves administering the test to a sample of examinees and calculating the average inter-item consistency
Kuder Richardson’s 20( kr-20)
- alternative to coefficient alpha
- can be used when test items are dichotomously scored
Split half reliability
- involves administering the test to a sample of examinees splitting the test in half, usually in terms of even and odd numbers and correlating the scores on the two halves
Problem with split half reliability
- You’re required to calculate reliability coefficient for two forms of the test that are half as long as the original test and shorter tests tend to be less reliable than longer ones
- It usually underestimates a test’s reliability and is usually corrected with the spearman Brown prophecy formula
Inter rater reliability
- important for measures that are subjectively scored
- provides info on the consistency of scores or ratings assigned by different raters
Percent agreement and cohens alpha coefficient
- methods for calculating inter rater reliability
Percent agreement
- can be calculated on two or more raters
- does not take chance agreement into account which can result in an ever overestimate of reliability
Cohen’s kappa coefficient
- aka the kappa statistic
- One of several inter-rator reliability coefficients that is corrected for chance agreement between raters
- used to assess the consistency of ratings assigned by two raters when the ratings represent a nominal scale
Factors that affect Reliability of subjective ratings
- can be affected by consensual observer drift
Consensual observer drift
- occurs when two or more raters communicate to each other while assigning ratings which results in an increased consistency but decreased accuracy in ratings and an overestimate of iterator reliability
Ways to reduce consensual observer drift
- not having raiders work together
- providing raiders with adequate training and regular monitoring the accuracy of raters ratings