Special Topic: The Replication Crisis Flashcards
(19 cards)
The Replication Crisis
● What is it?
● Why do we care?
● How bad is it?
● When you read papers, how do you know you can trust the findings?
Does learning about the replication crisis help?
● Undergraduate students who learned about replication for 1 hour…
○ Showed high endorsement that media attention was not a accurate indication of study reliability
○ Showed high agreement with suggestions about transparency and replication (e.g., publishing null
findings, making data open)
○ Understood the problems with flexible decision making in statistics
○ Showed slightly less trust in psychological findings
○ Showed greater appreciation for study design
○ Showed no decline in the desire to pursue graduate school
salmon experiment?
- a salmon was put in a fmri scan
- their was activated in the areas in the human brain when thinking about various social situations
- but the salmon was dead
- because - They did multiple tests with the guarantee that a % of them will be significant
False positive rate: % of results that will show as significant, tied to alpha
we see big effect sizes without replication to see if they indeed can be infared
The replication crisis:
Many psychology studies done in the past do not reproduce by modern standards.
- Effects nonexistent or in opposite direction
-
Previously significant results no longer significant
-they might be crerry picked or luck -
Some effects exist, many do not (before: many big effects
existed) - But why?
p hacking
- re-making the reserch until it fit into what the resercher wants
- get a lot of participants but at soon as it gets what they want they stop with the study
File drawer problem
- Doing lots of studies, knowing that alpha % of our studies (e.g. 5% for
alpha = .05) will show a false positive - Then, discarding the studies that didn’t support our hypothesis and
did not produce significant results - All that gets published is the study that happened to have significant results!
The rest remain in the file drawer
Significance incentives
- Once upon a time (and still today), journals only wanted to publish
significant findings
* Incentive to get published (and get a better chance at a job/tenure):
Publish significant findings
- But most findings in psychology are not significant
Why is this a problem?
- The studies we cite or build upon may show effects that are not
reliable or that don’t really exist - We lose our sense of the true state of cognitive processes
-
We are incentivized to discard studies that don’t show significant
results, leaving a lot of great science in the file drawer
What can we do about the replication crisis?
- Move beyond the p-value
- Be upfront about what you plan to do
- Be transparent in your work
- Move beyond the p-value
- your p-value should not be dancing around
- Effects in psychology are not single points; they are distributions
using distribution over p-values
- there should have a trashold
Using the distribution: Instead of looking at p-values..
- Compare confidence intervals
- Compare the distributions themselves, (how they overlap)
2. Be upfront about what you plan to do
its very easy
Pre-registration helps keep us accountable
- By reporting all studies and analyses you plan to conduct, insignificant
results can’t be put in the file drawer -
Incentive: some journals will promise to publish your results
regardless of their significance if you pre-register your analyses with
them
3. Be transparent in your work
- We want to replicate and see if our work can be replicated
- Replication only works when we know the conditions under which the original study took place!
Letting our studies replicated can feel scary…
-
Perspective: it is better, more honest science to let our work be
critically evaluated and tested by others - We build on each others’ work –
Be transparent in your work
How can we know what others did (and others, what we did?)
- Share your data
- Share your methods
- Be open to critiques/corrections if your work wasn’t replicated by
others
Data sharing
t’s easy to share your code, data, materials, etc. on a site like OSF
- Others can look at the raw data, try and replicate your analyses, and much more
What should I take away from this?
-
Many past studies in psychology cannot be replicated because:
-They were not fully transparent in their procedures and findings
-They just “happened” to be the one significant study out of many -
True effect sizes in psychology are smaller than we’ve been led to
believe
-There are fewer significant effects -
We can address these problems by:
-Incorporating all of your data into the analysis
-Being accountable to your research plans
-Being transparent with procedures and findings