Chapter 3: Survey research Flashcards
(22 cards)
Three ways to measure a variable with their problems
- Psychiological indicators: can be problematic because of confounding variables, measuring can be stressful, practical issue to test multiple people at once
- Questionnaires: biases, accuracy of responses
- Observations: differences between people, confirmation bias, behave differently when observed
Reliability
Refers to consistency
A necessary condition for construct validity, but it is not because there is a high reliability that there immediately is a good validity!
Three types of reliability: test-retest, interrator and internal
Test-retest reliability
We measure the variable at two measurement moments (same measure, same group of people)
High test-retest reliability: people have similar scores on both measures
Assumes stability in construct across measurement moments
Interrater reliability
At least two raters who observe behavior (observation studies or qualitative research)
Reliability of observations is measured by correlate (r) scores of both raters → consistent: high correlations between scores
Correlations can only be used on interval variables, when rating a nominal variable we use Cohen’s Kappa
Internal consistency reliability
Useful for multi-item scales measuring the same construct
We expect that items measuring the same construct correlate positively with each other → average inter-item correlation
Cronbach’s alpha to compute variable: the higher the score, the more reliable the results are (>0.70) → better alternative: omega
Two subjective ways to assess validity
- Face validity: does the measure appear to be a good indicator of the construct?
- Content validity: does the measure contain all relevant aspects of the construct?
Three empirical routes that use data to assess validity
- Convergent validity
- Discriminant validity
- Criterion validity
Convergent validity
You expect that there will be a (positive/negative) relationship between our measure and another variable we measured
How to measure: give people two questionnaires, evidence for convergent validity if the scores on both tests have a relationship with eachother
Discriminant validity
You expect that there will be no relationship between our measure and another variable we measured
We don’t expect people who score high on a measure to score high/low on another scale or measure (that has nothing to do with the construct we measure)
There is evidence for discriminant validity when there is no or a very weak correlation between both variables
Criterion validity
Is there evidence for a relationship between the measure and certain behaviors?
Can we predict how people would behave or how they would score on another measure based on this measure?
Makes use of the known-groups paradigm
Idea: think about a relationship of the variable with a certain outcome, collect data and check whether data supports expected pattern
Experience sampling methods
A representative sampling of immediate experiences in one’s natural environment
Singal-contingent sampling: send a notification through an app on phone and they have to react to it
Event-contingent sampling: fill out a survey when something happens
Strengths: high ecological validity, less sensitive to memory bias and common method bias, ideal to study changes in variables over time
Weaknesses: high participant burden → high drop-out, testing effects or reactivity
Four things to consider when developing a scale
- Choosing the question format
- Writing well-worded questions
- Encouraging accurate responses
- Validating the scale
Choosing the question format
Open questions: a question and a textbox where people write their answers
Closed question: answer options like dichotomous (yes/no), nominal (gender), ordinal (degree), interval
Four types of scales
- Likert scale: bipolar, indicate to which extent you agree, 6-7 answer options is recommended to eliminate neutral middle category
- Visual analogue scale: measuring pain level with ‘faces’ pain rating scale
- Semantic differential scale: scale with two endpoints that are opposites of each other, indicate where they would fall in between both points
- Guttman scale: statements and indicate whether or not they agree with them
Writing well-worded questions
Pay attention to:
Leading questions: formulated in a way that you automatically tend to choose a certain option
Double-barreled questions: questions that contain two questions and ask you to rate two different things at a time
Double negotiations: too complex
Order of questions: first questions might influence way people react to the next ones
Encouraging accurate answers
How much can we rely on self-reports: we tend to report more than we actually know + recollection bias
Respons bias
Four types of response biases
- Acquiescence: always selecting the same answer because you’re not paying attention
- Fence sitting: always choosing the middle/neutral option to not pick a side
- Social desirability: portraying a better picture of yourself by choosing the option you think is socially acceptable
- Faking bad (malingering): portraying yourself worse than you are
Three types of miscomprehensions
- Instructional miscomprehensions: not reading the instructions
- Sentential miscomprehensions: interpreting questions differently
- Lexical miscomprehensions: not understanding something in the question
Nine methods to detect ‘low quality data’
- Self-report items: ‘did you answer truthfully?’
- Bogus items: incorrect/impossible questions
- Instructed items: ‘please indicate …’
- Too fast response times
- Longstring: choosing the same answer options for nine times in a row
- Individual response variability
- Psychometric synonyms: opposite questions
- Personal reliability
- Mahalonobis D: scores that are very different from others → way to detect outliers
Observations
In a lot of situations, observations are better than self-reported data
Usually, there is a difference between hypothetical and real behavior in self-reports
Problems: observer bias, observer effect, reactivity
Observer bias
Observer effect
Reactivity
Observer bias: observers see what they expect to see and would like to see it confirmed
Observer effect: participants behave in line with expectations of the observer
Reactivity: participants react to being observed and change their behavior
Six ways to assure reliable and valid observations
- Unobtrusive observations: observing people who don’t know they’re being observed
- Wait it out: people can’t keep changing their behavior
- Measure the behavior’s result
- Use a codebook: describe what behavior you want to see
- Interrater reliability: use multiple observers
- Blind and double-blind design: participants don’t know the hypothesis / observer also doesn’t know hypothesis