Validity Flashcards
(35 cards)
What is the distinction between validity and reliability?
Validity = measurement of the test is valid when it measures what it is supposed to. Reliability = consistency of measurement
Why is it imprecise to say that a test is valid?
• Because validity refers to the measurement that is attained by use of the test
Construct underrepresentation:
Construct underrepresentation: aspects of the construct not assessed by the measure (maybe due to poor construction of the measure or changing knowledge of the construct)
Construct irrelevant variance:
variance in test scores related to things unrelated to the construct (e.g., random and systematic error). Reduces valid measurement.
Valid measurement:
overlap where measurement is measuring what it is supposed to
How does increased or decreased validity play into the relationship between these?
How does increased or decreased validity play into the relationship between these?
Decreased validity = more construct underrepresentation and/or more construct irrelevant variance.
Increased validity = less of those.
What is validity of measurement?
Already discussed above… maybe look into in more detail in the additional readings section.
overlap where measurement is measuring what it is supposed to… maybe look into in more detail in the additional readings section.
What are content and construct validity? How do these differ?
- Content validity: examination of aspects of the test itself. Confirmatory, non-statistical/non-quantifiable process of determining how well a measure overlaps with the construct.
- Construct validity: examination of relationship between test scores & scores of other measures
• Content validity only examines aspects of the test itself, whereas construct validity compares the test scores with scores on another measure of the same construct
Construct validity
Construct validity: examination of relationship between test scores & scores of other measures
Content validity
Content validity: examination of aspects of the test itself. Confirmatory, non-statistical/non-quantifiable process of determining how well a measure overlaps with the construct.
Content validity only examines aspects of the test itself, whereas construct validity compares the test scores with scores on another measure of the same construct
Determine the areas of the content domain that are measured by each test item
Boundary = identification of all sub-constructs that exist within that construct.
• Structure = relative weight or importance of each sub-construct within the larger construct (determined by examining research)
Determine the areas of the content domain that are measured by each test item
Making sure all sub-constructs are represented by items:
o To account for weight: adjust # of items or adjust values in scoring
o This avoids construct irrelevant variance
Compare the structure of the test with the structure of the content domain
Need to ensure your measurement reflects construct
o Increases content validity, decreases construct irrelevant variance & underrepresentation
Why is an assessment of content validity seen less in psych measurement than educational testing?
It is harder to assess content validity with psychological constructs. Our knowledge and conceptualization of the constructs change, making it difficult to identify boundary and structure.
What is face validity?
The degree to which a measure appears to measure the target construct.
Why is it good to have a measure that appears to be valid, but not at the expense of empirical validity?
High face validity can increase participants’ likelihood of continuing to respond. People hate wasting their time.
However, you may need low face validity to get honest answers for some constructs (e.g. racist attitudes).
What are the methods of determining construct validation?
- Correlational study: Correlations between measures of certain behaviors (either related or unrelated to our construct) and our test.
• Likelihood of 2 measures failing to tap into the same aspect of a construct is small.
• Easiest and cheapest option - Factor analysis: analyzing which groups of items “hang” together.
• Compares factor structure that results from a measure with the assumed factor structure of the construct.
• Gives a determination of which items conceptually intercorrelate/covary
• Hard to do and requires huge sample size - Experimental manipulation: manipulate the construct of interest (e.g., induce fear) and see if it relates to scores on your test.
• Best way EXCEPT problems:
o ethics (e.g., unethical to induce PTSD)
o only works with stable constructs
o more expensive and time-consuming
What is the difference between convergent and discriminant validity? What would you expect from each in terms of the resulting correlation analysis?
Convergent validity: we want significant correlations between test & measures of behaviors related to construct
Discriminant validity: we want non-significant correlations between our test & behaviors unrelated to construct
We expect a significant positive correlation when things are related (convergent validity).
We expect a nonsignificant correlation when things are not related (discriminant validity).
What types of information can be gained from use of the multitrait-multimethod matrix approach?
Note: You will need to determine and interpret this information from a matrix on the exam (see slides).
MTMM allows you to determine 3 things:
- Convergent validity
- Discriminant validity
- Check for method bias
What is method bias?
If there is a significant relationship (one or a couple) within the same method, it may indicate a training issue or a conceptual overlap between two traits.
What is a criterion?
A measure that could be used to determine the accuracy of a decision. Outcome measure.
What is criterion-related validity? (aka decision-making validity)
- Quantitative way to evaluate how successful decisions made from the test’s results are.
- We compare scores on our measure with scores on a criterion assessment (an outcome measure that helps you determine the accuracy of decisions).
- There is high criterion-related validity when there are significant correlations between our test and the criterion assessment.
- Assessed using either prediction validation strategies or concurrent validation strategies.
Why is a predictive validation strategy considered the ideal approach? Why isn’t it always used?
- Maximizes potential for covariance (avoids truncated range)
- BUT not always ethical or cost effective
How does one conduct a predictive validation procedure?
You screen everyone and then accept everyone (so you can see the full possible range in later assessment).
- Obtain test scores
- Obtain performance measures
- Correlate these with the test scores