Chapter 3 Issues in Personality Assessment Flashcards
(31 cards)
Assessment
Measuring of Personality
Observer Ratings
- Someone other than the person being assessed
- Interviews
- Watching his or her actions
- Observe a person’s belongings
Interviews
- Can be especially insightful if person is asked to talk about something other than themselves- boundaries come down
- In the moment observation vs. summative judgments
Snoopology
- studied people’s offices, bedrooms, and other personal domains
- People “portray and betray” their personalities by the objects and mementos they surround themselves with. (Gosling, 2008)
- Identity claims: symbolic statements about who we are
- Indicators of how we want to be regarded
- Can be directed to other people who enter our space, or they can be directed to ourselves, reminders to ourselves of who we are
- Feeling regulators: help us manage our emotions
- Behavioral residue: physical traces left in our surrounds by everyday actions
- Tell more the residue, the less organized you probably are.
- Give an indication of what sorts of things take place repeatedly in your life space
Self Reports
- People themselves indicate what they think they’re like or how they feel or act.
- Introspection
- Ask people to respond to a specific set of items
- Many formats
- True-false
- Multipoint rating scale
- Some focus on a single quality of personality
Inventory
- Measure that assesses several dimensions of personality
- Go through each step of development for each scale of the inventory, rather than just one.
- Multiple scales
Implicit Assessments
*Attempt to find out what a person is like from the person (like self-reports) but not by asking him or her directly.
*Discover people’s unconscious attitudes or perceptions- often include parts of their personalities they may be ignorant of or try to hide
-Given a task of some sort that involves making judgments about stimuli.
-The pattern of responses (ex: reaction times) can inform the assessor about what the person is like
*Ex: Implicit Association Test (IAT)
-Semantic properties in memory that are believed to be hard to detect by introspection
-Categorize a long series of stimuli as quickly as you can
Motive approach to personality
*The person being assessed produces a sample of “behavior”
-Action
-Internal behavior-heart rate
-Answering questions
Subjective Measures
*Interpretation is part of the measure
Objective Measures
*The measure is of a concrete physical reality that requires no interpretation
Reliability
- Once you’ve made an observation about someone, how confident can you be that if you looked again a second or third time you’d see about the same thing?
- Consistency/repeatability
Error
- randomness
- Can be reduced, but not eliminated
- Repeat the measure- make the observation more than once
- Measuring the same quality from a slightly different angle or using a slightly different “measuring device”
Internal Consistency/Internal Reliability
- Each observer or item carries its own error, so to cancel out, we use many different question items. Internal Consistency refers to the extent that they all agree with one another.
- Within a set of observations of a single aspect of personality
- The items are highly reliable means that people’s responses to the items are highly correlated.
- More items in self-reports
- Different telescrope
- Different math problem
Split Half Reliability
- Separate the items into two subsets (odd vs. even-numbered items), add up people’s scores for each subset and correlate the two subtotals with each other
- If the two halves of the item set measure the same quality, people who score high on one half should also score high on the other half, and people who score low on one half should also score low on the other half.
- Measure internal consistency
Inter-rater Reliability
- both see about the same thing when they look at the same event
- Trained in how to observe what they’re observing
- Agreement among different raters/observers
Item response theory (IRT)
- Attempt to increase the efficiency of assessment, while reducing the number of items
- Determining the most useful items, and the most useful response choices, for the concept being measured
- Creation of response curves: show how frequently each response is used, and whether each choice is measuring something different from other choices
- Determines the “difficulty” of an item
- Computerized adaptive testing (CAT)
Computerized adaptive testing (CAT)
*Ensures that less difficult items are not given after an item of medium difficulty has been endorsed
Stability Across Time
Extent to which measurements are stable over time
*Test Re-Test Reliability
Test Re-Test Reliability
*Giving the test to the same people at two different times
Reliability concerns repeatability across time
Validity
*Whether what you’re measuring is what you think you’re measuring (or what you’re trying to measure).
Construct Validity
- How well deos operational definition (the event) match the conceptual definition (the abstract quality you have in mind to measure)?
- Close, high validity; aren’t close, low validity
- Other subtypes of validity all provide support for construct validity
Conceptual and Operational Definition
- Ex: Love
- Conceptual definition: a strong affection for another person
- Operational definition: ask the person you’re assessing to indicate on a rating scale how much she loves someone/measure how willing she is to give up events she enjoys in order to be with him.
Criterion Validity (Predictor Validity)
- Other manifestations (behavioral index/external criterion) of the quality it’s supposed to measure
- Does your measurement produce results that are consistent with an external criterion, such as a trained observer?
- Does it accurately predict what it purports to measure?
- Tests how well the measure predicts something else it’s supposed to predict
- Often relevant when developing a new scale- if it’s consistent with other, already established measurements of same construct
- Best way to predict construct validity
- Too often, researchers choose criterion measures that the poor reflections of the construct.
Convergent Validity
- Showing that the measure relates to characterisitics that are similar to, but not the same as, what it’s supposed to measure.
- The evidence “converges” on the construct you’re interested in, even though any angle finding by itself won’t clearly reflect the construct.
- Ex: Measure dominance
- Should relate at least a little to measures of qualities such as leadership (positively) or shyness (inversely)
Discriminant Validity
- It does not measure qualities it’s not intended to measure- especially qualities that don’t fit your conceptual definition of the construct.
- Defense against the third-variable problem