L3: Individual Differences & validation Flashcards
define interpersonal skills
skills related to social sensitivity, relationship building, working w others, listening, and communication
define incremental validity
how much a new test or predictor adds value byeond what is already measured by existing predictors.
- empirical
aka it shows whether adding a new test (like SJTs) improves the ability to predict an outcome (job performance- compared to traditional tests (cognitive ability tests)
why do SJTs predict long-term success?
this is important for its validity as a predictor for future success
- they assess soft skills like communication, empathy, and decision making
- these skills remain stable over time & are crucial for job performance
define general procedural knowledge
the knowledge someboy has acquired about effective & ineffectives courses of trait related behaviour in situations like those described in the SJT
in LIevens study, general procedural knowledge reltaes to students procedural knowledge about (in)effective behaviour in interpersonal situations (with patients) as depicted in the SJT items
what are the implications of LIevens study?
- SJT are valuable for student seleciton alongside cognitvie tests
- predicting post academic success is important, not just academic success
- long term validity of SJTs supports their use in professional admissions
define reliability
consistency of a test or measurement. so whether the measure is:
- dependable
- stable
- consistent over time
- gives the truest picture of someone’s abilities/characteristics
how do you test reliability?
- correlation coefficient methods
- test retest, coeff of stabiility
- parallel/alternate forms, coeff of equivalence
- coeff of stability & equivalence: combines different sources of error
- internal consistency (alpha)
- interrater reliability
define correlation coefficient
degree of consistency/agreement between 2 sets of independently derived scores (r)
define internal consistency
degree to which the items in one test are intercorrelated
split half: split test into 2 equivalent halves
chronbachs alpha: mean of all possible split half
most commonly used
ex: all math problems in a test should measure math ability
define interrater reliabilty
agreement between raters on their rating of some dimensiion
ex: olympic judges giving similar scores for a gymnast
what is parallel forms reliability? and example?
consistency across different versions of the test
ex: 2 versions of the SAT should give similar results
what is test retest reliablity & an example
consistency over time
ex: taking an IQ test twice & getting similar scores
what is a “good” reliability?
- depends on what you want to do with the scores
- the more important the decision (like life or death for reliability of fMRI scan) the more precise the measure needs to be. generally for research in social sciences >.7 and for selection decisiosn >.9
what does validity look at?
whether a measure is:
- actually measuring the construct its supposed to measure &
- whether the decisions based on that measure are correct
define content validity
The extent to which test items cover the intended
performance domain
how do you measure content validity?
- rational examination of the manner in which the performance domain is sampled by the predictor
- SMEs: degree of agreement among them regarding how essential a particular item/test is (if more than 1/2 SMEs say an item is essential, that item has at least some content validity, CVI formula)
- difficult for more abstract constructs
define criterion validity
- How well a test predicts real-world outcomes.
- empirical relationship between predictor & criterion (performance measure)
- subtypes: predictive & concurrent
what is predictive criterion validity vs concurrent criterion validity?
predictive: measures how well a test predicts future success
(test scores collected before the criterion like college entrance exam)
concurrent: measures how well a test correlates w current performance (test scores & criterion data collected at the same time like job skills test)
how do you measure predictive validity?
- use stats to demonstrate the actual relationship between predictors & criteria
- for ex: linear realtionship Y = a + bX + c
1. measure candidates on predictor during selection (like conscientiousness)
2. select candidates without using the results (need to validate first)
3. obtain measurement of criterion performance later -> time period depends on type of job & how much training is needed (approx 6m)
4. assess the strength of the relationship (stats)
how do you measure concurrent validity?
both predictor & criterion data gathered from same employees (incumbents)
- cross sectional study
what are some issues affecting validity?
- range restriction
define meta analysis
combines multiple studies to determine overall validity
what are the pros & cons of meta analysis
Pros: Resolves inconsistencies, finds general patterns.
Cons: Affected by publication bias (only strong results get published), study differences, and data quality.
what is range restriction?
if only a specific group is tested, the validity appears lower than it really is
ex: if a study only includes top students, the entrance test might seem weak at predicting performance