CHAPTER 6: VALIDITY Flashcards

1
Q

It is a judgment or estimate of how well a test measures what it purports to measure in a particular context.

A

Validity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

It is a logical result or deduction.

A

Inference

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

It is the process of gathering and evaluating evidence about validity.

A

Validation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Studies conducted to gather evidence that a test measures what it is intended to measure and is appropriate for its intended use.

A

Validation Studies

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Validation studies done by the test user on their own group of test takers to ensure the test is valid for that specific population or setting, especially when changes to the test or its administration are made.

A

Local Validation Studies

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

This is a measure of validity based on an evaluation of the subjects, topics, or content covered by the items in the test, and ensures that the test covers the full content of the construct it aims to measure.

A

Content Validity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

This refers to how well one measure predicts or correlates with an outcome or another measure (the “criterion”). This is a measure of validity obtained by evaluating the relationship of scores obtained on the test to scores on other tests or measures.

A

Criterion-related Validity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

This refers to whether the test truly measures the psychological concept or construct it intends to measure, like intelligence, depression, or motivation. This is a measure of validity that is arrived at by executing a comprehensive analysis of:

a. how scores on the test relate to other test scores and measures, and;
b. how scores on the test can be understood within some theoretical framework for understanding the construct that the test was designed to measure.

A

Construct Validity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

A traditional framework of test validity that includes three main types: content validity, criterion-related validity, and construct validity, where construct validity serves as the overarching or “umbrella” concept.

A

Trinitarian View

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

The extent to which a test accurately measures a behavior, thought, or emotion in the real-world setting or situation where it naturally occurs.

A

Ecological Validity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

It refers to the degree to which a test appears to measure what it claims to measure, based on surface-level judgment. It relates more to what a test appears to measure to the person being tested than to what the test actually measures.

A

Face Validity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

A detailed plan that outlines the content areas, number of items for each area, and the structure and organization of a test.

A

Test Blueprint

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

This type assesses whether a test correlates with a criterion measured at the same time. It is an index of the degree to which a test score is related to some criterion measure obtained at the same time (concurrently).

A

Concurrent Validity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

This type evaluates how well a test predicts a future criterion. It is an index of the degree to which a test score predicts some criterion measure.

A

Predictive Validity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

It is a standard used to make a judgment or decision, more specifically referring to the benchmark against which a test or test score is evaluated.

A

Criterion

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

A characteristics of criterion wherein it is pertinent or applicable to the matter at hand.

A

Relevant

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

A characteristics of criterion wherein it is valid for the purpose for which it is being used.

A

Valid

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

A characteristics of criterion wherein it means that the criterion is free from influence by the predictor measures used in the test, ensuring that the standard against which a test is evaluated is not biased or artificially aligned with the test itself.

A

Uncontaminated

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

It is the term applied to a criterion measure that has been based, at least in part, on predictor measures.

A

Criterion contamination

20
Q

It is the extent to which a particular trait, behavior, characteristic, or attribute exists in the population (expressed as a proportion).

21
Q

It may be defined as the proportion of people a test accurately identifies as possessing or exhibiting a particular trait, behavior, characteristic, or attribute.

22
Q

It may be defined as the proportion of people the test fails to identify as having, or not having, a particular characteristic or attribute.

23
Q

It is a miss wherein the test predicted that the testtaker did possess the particular characteristic or attribute being measured when in fact the testtaker did not.

A

False Positive

24
Q

It is a correlation coefficient that provides a measure of the relationship between test scores and scores on the criterion measure.

A

Validity Coefficient

25
It is defined here as the degree to which an additional predictor explains something about the criterion measure that is not explained by predictors already in use.
Incremental Validity
26
It is an informed, scientific idea developed or hypothesized to describe or explain behavior. These are unobservable, presupposed (underlying) traits that a test developer may invoke to describe test behavior or criterion performance.
Construct
27
One of the evidences of construct that refers to how consistently the items on a test measure the same construct. A test is said to be homogeneous when all its items correlate positively and significantly with the total test score, indicating that they are likely measuring the same underlying trait or ability.
Evidence of Homogeneity
28
It refers to one type of construct validity evidence that examines whether a test reflects expected developmental trends over time. If a construct is known to change with age, then valid test scores measuring that construct should also reflect such changes.
Evidence of Changes with Age
29
It is a form of construct validity evidence that examines whether a test can detect meaningful score changes due to an intervening experience or treatment. If test scores shift in expected directions after a known intervention, this supports the test’s validity in measuring the intended construct.
Evidence of Pretest–Posttest Changes
30
It is a form of construct validity evidence that is also known as method of contrasted groups, and is a way to support construct validity by showing that test scores differ predictably between groups known to differ on the construct.
Evidence from Distinct Groups
31
If a new test measures what it claims to (e.g., test anxiety), it should show strong positive correlations with established tests of the same construct and moderate correlations with related constructs (e.g., general anxiety). High correlations provide convergent validity evidence.
Convergent Evidence
32
If a test is valid, it should not strongly correlate with unrelated constructs. For example, the Marital Satisfaction Scale (MSS) was found to have no significant correlation with a Social Desirability Scale, showing that MSS responses weren’t just socially desirable answers. This supports discriminant validity.
Discriminant Evidence
33
It is a type of construct validity evidence that shows low or no correlation between a test and variables it should not theoretically be related to.
Discriminant evidence (or discriminant validity)
34
It is a statistical technique introduced by Campbell & Fiske (1959) used to assess the construct validity of a set of tests by examining both convergent and discriminant validity. It involves measuring two or more traits using two or more methods, then comparing the correlation patterns.
Multitrait-Multimethod Matrix (MTMM)
35
It is a statistical technique used to identify underlying factors or dimensions that explain the pattern of correlations among test items or subtests. It helps determine whether a test actually measures the constructs it claims to measure.
Factor Analysis
36
It refers to a systematic error in a test that distorts the measurement of a psychological or educational trait for a particular group. It leads to unfair advantages or disadvantages not due to actual differences in the trait being measured, but due to irrelevant factors embedded in the test.
Test Bias
37
For psychometricians, it is a factor inherent in a test that systematically prevents accurate, impartial measurement.
Bias
38
It is a judgment resulting from the intentional or unintentional misuse of a rating scale.
Rating Error
39
It is a type of rating error that is also known as generosity error, where the rater consistently gives higher scores than what is objectively warranted. This happens due to a tendency to be overly generous, rather than strictly evaluating actual performance or behavior.
Leniency Error
40
It is a rating error in which the rater consistently gives lower or harsher scores than what is justified by the actual performance or behavior. It reflects a tendency to be overly critical.
Severity Error
41
It is a type of rating error where the rater avoids giving very high or very low scores, and instead clusters all ratings around the midpoint of the scale, regardless of actual performance.
Central Tendency Error
42
It is a procedure that requires the rater to measure individuals against one another instead of against an absolute scale.
Rankings
43
It may be defined as a tendency to give a particular ratee a higher rating than he or she objectively deserves because of the rater’s failure to discriminate among conceptually distinct and potentially independent aspects of a ratee’s behavior.
Halo Effect
44
It refers to the impartial, just, and equitable use of a test across individuals or groups.
Test Fairness
45
It is value-driven and deals with ethical, legal, and social considerations.
Fairness