Statistics Flashcards
(33 cards)
What is a t-test and why would you use it?
a t-test is a statistical test that is used to compare the means of two groups.
it is used to determine if populations have significant difference or this difference is down to variance.
What is a P-Value?
the probability that the results from your sample data occurred by chance.
What P-value should you use and why?
The p-value threshold of 0.05 is widely used because it represents a practical balance between the risks of Type I and Type II errors.
It is essential because it helps you make informed decisions about your hypotheses based on the significance of your results.
P value > 0.05 is the propability that the null hypothesis is true.
What is the T-value?
The t-value is a statistic used in hypothesis testing to measure the difference between the observed sample mean and the expected population mean, accounting for the variation in the sample data.
A higher magnitude of the t-value indicates stronger evidence against the null hypothesis (H0), suggesting that the observed difference is more significant and less likely to be due to random chance.
Therefore, the greater the t-value, the more compelling the argument that the null hypothesis may not be true.
T critical at 0.05 p is 1.895 - so, we accept T values above this.
You complete a T test comparing two variables, your T value is 0.75, what does this mean and why?
this t value is below t critical of 1.895 so this means that there is no significant difference between the groups, this is because the t critical at 0.05 P is 1.895, so anything below would mean the null hypothesis likelihood is more than 5%.
If a T-test comparing test scores between two study methods gives a T value of 2.8, is there a significant difference? Why?
Yes, a T value of 2.8 suggests a significant difference, as it indicates a notable difference between group means compared to within-group variation, likely rejecting the null hypothesis.
How do you calculate degrees of freedom
df = N1 + N2 - 2
N = sample number
What is degrees of freedom?
Degrees of freedom (df) are the number of values in a calcution that are free to vary.
It is the flexibility you have when working with a dataset.
This is often defined as sample size - 1 (n-1).
The more df (larger sample) means a more reliable estimate of population parameters as there’s more data to reduce uncertainty.
Fewer df (smaller sample) means higher uncertainty in estimates, requiring adjustments to the t -distribution to account for this.
For two independent samples, the formula is:
df = (n-1)+(n+1)
where n and n are the sample sizes.
What is a paired t test?
A paired T-test compares the means of two related groups, such as measurements from the same subjects at two different times or conditions, to determine if there is a significant difference.
It calculates the mean of the differences and tests if this mean is significantly different from zero.
What is a chi-squared test
The chi-squared test is a statistical test used to see if there’s a meaningful difference between what we expect to find in data and what we actually observe. It’s particularly useful when working with data divided into categories (like types of fruits, survey answers, etc.).
It is used when you have categorical data and want to see if the differences between groups are just by chance or if they’re statistically significant.
When should you do a two-tailed test?
A two-tailed test is used to detect differences in both directions (positive or negative) between groups or conditions.
It tests for any significant difference, not just an increase or decrease, making it more flexible and applicable when there’s no specific prediction about the direction of the effect.
What is beta in power calculations?
beta (β) is the probability of making a Type II error (a false negative), which is failing to reject the null hypothesis when there is a true effect.
It represents the chance of missing a real difference. Typically, power = 1 - β, with a lower beta indicating higher test sensitivity.
What is alpha in power calculation?
alpha (α) is the probability of making a Type I error, which occurs when the null hypothesis is incorrectly rejected (false positive).
It represents the significance level of a test, commonly set at 0.05, meaning there’s a 5% chance of finding a significant result by chance.
When would you use Anova?
ANOVA (Analysis of Variance) is used when comparing the means of three or more groups to determine if at least one group mean is significantly different from the others.
It’s helpful in experiments with multiple groups or conditions, testing for overall differences rather than just pairwise comparisons.
What is quantitive data?
Quantitative data is numerical information that represents measurable quantities.
It includes data that can be counted or measured, such as height, weight, age, or test scores, and is typically analysed with statistical methods to identify patterns or relationships.
What is qualitative data?
Qualitative data is non-numerical information that describes characteristics, qualities, or categories.
It includes data like interview responses, observations, and descriptions, often used to understand concepts, opinions, or experiences.
What is an F value?
An F value is a statistic calculated in ANOVA tests, comparing the variance between group means to the variance within groups.
A higher F value indicates a greater likelihood that there is a significant difference among group means.
It’s used to assess whether the observed variation is due to group differences or random chance.
What is effect size?
Effect size is a quantitative measure that reflects the magnitude of a treatment or intervention’s effect in a statistical analysis.
It provides a standardised way to compare results across different studies or experiments, allowing researchers to understand the practical significance of their findings beyond just statistical significance.
What is a Type I error?
A Type I error occurs when the null hypothesis is incorrectly rejected, meaning a false positive result.
It suggests finding a significant effect when there is none.
The probability of a Type I error is represented by alpha (α), often set at 0.05, indicating a 5% chance of mistakenly declaring a result significant due to random chance.
What is a power calculation, and what does it tell you?
A power calculation determines the likelihood that a test will detect a true effect if one exists, quantified as statistical power.
It calculates the probability of avoiding a Type II error (false negative).
Power is affected by sample size, effect size, significance level (alpha), and variability.
Higher power (typically 80% or above) indicates a stronger ability to detect meaningful differences, ensuring that the test is sensitive enough to find real effects.
Why do we do statistical analysis?
Statistical analysis is conducted to interpret data objectively and identify patterns or relationships.
It helps test hypotheses, assess the significance of results, and calculate power to ensure the analysis is sensitive enough to detect true effects.
This process controls for chance, provides reliable conclusions, and supports generalizable findings.
What does an a priori power calculation measure, and why would you use this?
An a priori power calculation estimates the sample size needed to achieve sufficient power (typically 80% or higher) before conducting a study. It ensures the study is designed with enough participants to detect a true effect if one exists, reducing the risk of a Type II error (false negative). This calculation is essential for planning reliable, efficient studies that are statistically valid.
What does a post-hoc power calculation measure, and why would you use this?
A post-hoc power calculation measures the power of a study after data collection and analysis to determine the likelihood that the study could detect an effect if one exists. It’s used to assess whether a non-significant result might be due to insufficient power or a true absence of effect, helping to evaluate the robustness and reliability of study findings retrospectively.
What is a manipulative experiment?
A manipulative experiment is a study in which researchers actively alter or manipulate one or more independent variables to observe the effects on a dependent variable. This type of experiment allows for testing causality, as it controls other variables to isolate the impact of the manipulated factor(s). It’s commonly used in laboratory and field research to establish cause-and-effect relationships.