Week 5 - Open Data, Reproducibility and Replicability Flashcards
(42 cards)
What is verifiability
A statement is meaningful when it can be verified empirically
What is the induction problem
To establish a law like “All swans are white” we must observe all swans
What is falsifiability
a statement is a valid theory if it makes predictions that can be tested, and can be falsified by a counterexample
What is a non-falsifiable theory
“One day there will be a human that can breathe underwater”
What is a Desideratum
in empirical work, we want to connect observations/measuerments with a falsifiable hypothesis or theory
What is Hypothesis testing
We need to establish a null hypothesis and an alternative hypothesis for falsifiability
What is transparency
In the ideal world, a study is fully transparent, in terms of what hypothesis is being challanged, what methodology was used, and what results were obtained.
What is Open data
All data should be available in order for other researchers to evaluate the study, or reuse materials
Open data is required or not required?
required in some academic journals
What could be the reasons why data is not shared?
- No time
- NO access
- Privacy
- Propriotery data (companies dont want to share their data)
how to combat data not being shared (no open data)
enforcing open data as a journal, peer review practice
Open data is necessary but not sufficient to guarantee good research
What is replicaiblity
The ability of a researcher to duplicate the results of a prior study if the same procedures are followed but new data is collected.
What is reproducibility
The ability of a reasearher to duplicate the results of a prior study using the same maternals as were used by the original investigator.
What are research artifacts
Any concrete object that was used in the execution of a study and that is needed to reproduce the study. Examples:
* Paper/report
* Dataset
* Model
* Software
What is a taxonomy of best practices on paper/report
peer review and checklists
What is the taxonomy of best practices of the dataset
Data annotation
What is the best taxonomy of best practices of Model and Software
ML best practices
What are the problems we can encounter when doing ML paper
- The data: the way we set up our data into a splits may impact performance
- The network: many current models are DNN, meaning they consist of many layers with low interpretability
How can we combat that. Give 2 methods
- Reproducibility from within: things researchers can do to increase the quality of their research.
- Reproducibility from outside: things reviewers should pay attention to
What is generalization
your model’s ability to adapt properly to new, previously unseen data, drawn from the same distribution as the one used to create the model
What is overfitting
The model performs very well on the training data but poorly on the validation data.
what is underfitting
the model doesn’t perform well anywhere
How to prevent loss hacking
Require authors to include loss statistics
how to combat underspecification (not enough detail and not reproducible)
Provide the selected train,val,test set and ideally the code that we used to create the split. This makes sure that the report is reproducible