Stats Flashcards

(56 cards)

1
Q

Null hypothesis (H0)

A

No difference between the two groups (default hypothesis)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Alpha (type I) error

A

Incorrect rejection of the null hypothesis - conclude there is a difference when there is no true difference

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Beta (type II) error

A

Incorrect acceptance of the null hypothesis - conclude there is no difference when there is

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Standard deviation

A

Measure of the variability of the data from the mean, equal to the square root of the variance

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Standard error

A

Measures the variability of means when many similar samples were taken from the population of possible measurements - how close the sample mean is likely to be to the true population mean

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Power

A

The ability to detect a significant result where a true difference exists (the probability of rejecting the null hypothesis when the alternative hypothesis is true)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Accuracy

A

How close a value is to the true value

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Precision

A

How consistent results are when measurements are repeated

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Validity

A

Suitability of the experimental method to address the aim of the experiment

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Internal validity

A

If outcome is related to aim

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

External validity

A

If results are applicable to the wider population

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Odds ratio

A

How much more likely one group is to be in an outcome group than another

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What is the alpha (type I) error also known as?

A

Specificity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What are alpha (type I) errors directly related to (statistically)?

A

P value

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What are beta (type II) errors also known as?

A

Sensitivity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What is the equation for calculating power (in simple terms)?

A

1-beta error

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

What power value do we typically want an experiment to have?

A

Between 0.8 and 0.9

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

What does power relate to (statistically)?

A

Sample size and effect size. If sample size is large enough, even a tiny effect size may be statistically significant (but not necessarily clinically or functionally significant)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

When may precision and accuracy become uncoupled?

A

Systematic error

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

When is mean vs median usually used?

A

Mean - when data are normally distributed with no major outliers

Median - when data are skewed or non-normally distributed

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

If the median of data is used, what should be used to measure spread?

A

Interquartile range

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

If the mean is used, what should be used to measure spread?

A

Standard deviation - spread of direct data

Standard error - where the real mean may be based on the population

X% confidence intervals - the range containing the population mean X% of the time based on the data collected

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

How can we test for normality?

A

Visually of q-q plots

Using tests such as the Shapiro-Wilk test

24
Q

If data are normally distributed, what type of test is used?

25
If data are not normally distributed, what type of test is used?
Non-parametric tests
26
What do non-parametric tests typically do?
Rank data and analyse whether a difference exists in the average rank between groups
27
Why is ANOVA preferable to multiple t tests?
Reduced risk of type I errors which would occur due to multiple comparisons
28
What does ANOVA tell you? What does this mean?
That a difference exists between any of the groups tested, but not where. One must then use post-hoc tests to compare means and identify where the difference is (these will include corrections)
29
What does the Tukey test do?
Compares
30
What does the Tukey test do?
Compares the differences in means between all groups
31
What does the Dunnett correction do? When may this be used?
Compares the mean of each group to a control value - could use when samples are taken periodically and you want to test when they become significant (known boundary)
32
What does the Bonferroni correction do? How?
It adjusts the p value for multiple testing - p is changed to be equal to alpha divided by the number of tests performed
33
What are the 5 main assumptions of T tests?
Data are normal Groups are independent (unpaired)/ samples are independent (paired) Equal variance Measurements are continuous No significant outliers
34
What are the 4 main assumptions in ANOVAs?
Data are normal Variance is equal Samples are independent Observations within each sample are random
35
What are the 4 assumptions of Pearson's test?
Continuous variables Paired values for each observation No outliers for either variable Data are linear
36
What information is required to calculate the number of replicates required to detect a known true difference?
The variation and size of difference expected, the accepted alpha/beta threshold - use past literature
37
When shouldn't non-parametric tests be used?
When n is very small i.e. below 6
38
What is the median?
Value in the middle when listing them all. ## Footnote Great for non-normally distributed data.
39
What is the mean?
Average of all values. ## Footnote Skewered by outliers.
40
What is a hazard ratio?
Treatment hazard rate divided by placebo hazard rate. ## Footnote E.g. 10% in treatment die, 20% in control die -> hazard ratio = 0.5 '50% decrease in deaths'. E.g. Hazard ratio = 0.64 -> 36% decrease in deaths in treatment compared to control.
41
What is a confidence interval?
We can be 95% sure that the true number lies within that value.
42
What are the two types of significance?
Statistical and clinical.
43
What is statistical significance?
P value is less than 0.05. ## Footnote Note a given hazard ratio has a 0.005 p value then there is a 0.5% chance of this hazard ratio occurring despite no clinical difference.
44
What is clinical significance?
Depends on the application and effect size. ## Footnote A statistically significant hazard ratio of 0.98 is only a 2% decrease in death. In clinic this could or could not be significant depending on how many people are affected. Another example: 1 mmHg reduction in BP with antihypertensive (that might even have side effects) is probably not significant in clinic.
45
What are parametric tests?
Deal with normally distributed data. ## Footnote More likely to get a statistically significant result, usually higher power.
46
What are non-parametric tests?
Deal with non-normal data. ## Footnote Either do a t-test to show difference between two groups (that is a parametric test?) Or transform data e.g. log transform, other exponents to make it normally distributed and use parametric tests.
47
What is a technical replicate?
Taking the same reading multiple times on the same sample. ## Footnote E.g. BP of one individual three times to make sure measurement is correct.
48
What is a biological replicate?
Same treatment to different samples. ## Footnote E.g. take BP of six different people.
49
What are error bars?
Could be standard deviation, standard error, or 95% confidence interval. ## Footnote Standard error is commonly used over the standard deviation as its SD divided by sample size and thus makes the error bars look smaller.
50
What is a test of multiple comparison, why is it important, and what is post-hoc testing?
Importance: p = 0.05 -> 1 in 20 tests will give a false positive. ## Footnote Thus do a test for multiple comparisons such as ANOVA. Post-hoc testing e.g. Bonferroni is then done to see differences between individual samples and a control or between all possible combinations - some tests have higher false positives other false negatives.
51
What are the conditions to be able to run a t-test and what is the result?
Data normally distributed, equal variance in both groups of samples, data is continuous. ## Footnote Result: one p-value - whether or not there is a statistical significance between the two groups.
52
What is the difference between one-way ANOVA and two-way ANOVA?
One way: multiple sample groups with only one variable e.g. effect of ramipril, amlodipine, both or placebo on BP. ## Footnote Two way: multiple sample groups with multiple variables e.g. ramipril, amlodipine, placebo effect on BP in old and young people or in high vs low doses.
53
What is an ANOVA readout?
A single p value that tells you whether or not there is some difference between some of the data that you are testing.
54
What do we do after an ANOVA?
Post hoc testing. ## Footnote E.g. run a t-test on each of the pairs with a correction such as Bonferroni.
55
What is a Bonferroni correction?
If you’re doing five post-hoc tests you have to divide your p value by the number of tests. ## Footnote E.g. 5 -> p value of each post hoc test would have to be <0.01 for the result to be statistically significant.
56
What is a power analysis?
Done before a study and takes into account: clinical effect expected, variance expected, statistical tests to be run, what acceptable false positive is. ## Footnote And computer answers you: which sample size is needed to make sure test will be significant and avoid false negatives.