Ch8 - Confidence Intervals, Effect Size, and Statistical Power Flashcards
What are the new statistics?
- Effect sizes
- Confidence intervals
- Meta-analysis
Point estimate
Confidence Intervals
- A summary statistic from a sample that is just one number used as an estimate of the population parameter - “best guess”
- The true population mean is unknown - and we take a sample from the population to estimate the population mean
- EX: In studies on gender differences in math performance - the mean for boys, the mean for girls, and the difference between them, are point estimates
Interval estimate
Confidence Intervals
- Based on a sample statistic and provides a range of plausible values for the population parameter
- Frequently used by the media, often when reporting political polls, and are usually constructed by adding and subtracting a margin of error from a point estimate
What is the interval estimate composed of (EQUATION)?
Confidence Intervals
interval estimate = percentage, + and - the margin of error
Confidence intervals: we’re not saying that we’re confident that the population mean falls in the interval, but rather…
Confidence Intervals
we are merely saying that we expect to find the population mean within a certain interval a certain percentage of the time - usually 95% - when we conduct this same study with the same sample size
Confidence level vs. interval:
Confidence Intervals
- Level - the %
- Interval - range between the two values that suround the sample mean
Calculating confidence intervals with distributions
Confidence Intervals
- Draw a normal curve that has the sample mean at its center (NOTE: different from curve drawn for z test, where we had population mean at the center)
- Indicate the bounds of the confidence interval on the drawing
- Determine the z statistics that fall at each line marking the middle of 95%
- Turn the z statistics back into raw means
- Check that the confidence interval makes sense
Step 1 to calculating CI
Confidence Intervals
Draw a normal curve that has the sample mean at its center (NOTE: different from curve drawn for z test, where we had population mean at the center)
Step 2 to calculating CI
Confidence Intervals
- ** 2: Indicate the bounds of the confidence interval on the drawing**
- Draw a vertical line from the mean to the top of the curve
- For a 95% confidence interval we also draw two small vertical lines to indicate the middle 95% of the normal curve (2.5% in each tail, for a total of 5%)
- The curve is symmetric, so half of the 95% falls above and half falls below the mean
- Half of 95% = 47.5%, represented in the segments on either sides of the mean
Step 3 to calculating CI
Confidence Intervals
3. Determine the z statistics that fall at each line marking the middle of 95%
- To do so: turn back to the z table
- The % between the mean and each of the scores is 47.5% - when we look up this % in the z table, we find a statistic of 1.96
- Can now add the z statistics of -1.96 and 1.96 to the curve
Step 4 to calculating CI
Confidence Intervals
4. Turn the z statistics back into raw means
- Need to identify appropriate mean and SD to use formula
- Two important points to remember:
- Center the interval around the sample mean (not the population mean), so use the sample mean in the calculation
- Because we have a sample mean (rather than an individual score), we use a distribution of means - so we calculate standard error as the measure of spread:
Step 5 to calculating CI
Confidence Intervals
5. Check that the confidence interval makes sense
* The sample mean should fall exactly in the middle of the two ends of the interval
Statistically significant doesn’t/does mean…
- Does NOT mean that the findings from a study represent a meaningful difference
- ONLY means that those findings are unlikely to occur, in fact, if the null hypothesis is true
How does an increase in sample size affect SD and the test statistic? What dooes this cause?
The effect of sample size on statistical significance
- Each time we increased the sample size, the SD decreased and the test statistic increased
- Because of this, a small difference might not be statistically significant with a small sample but might be statistically significant with a large sample
Why would a large sample allow us to reject the null hypothesis than a small sample? (EXAMPLE)
If we randomly selected 5 women and they had a mean score well above the OkCupid average, we might say “it could be chance”; but if we randomly selected 1000 women with a mean rating well above the OkCupid average, it’s very unlikely that we just happened to choose 1000 people with high scores
Effect size
- Indicates the size of a difference and is unaffected by sample size
- Can tell us whether a statistically significant difference might also be an important difference
- Tells us how much two populations DO NOT overlap - the less overlap, the bigger the effect size
- DECREASING OVERLAP IS IDEAL!
How can the amount of overlap between two distributions be decreased? TWO WAYS:
1: overlap decreases and effect size increases when means are farther apart (distance wise)
2: overlap decreases and effect size increases when variability within each distribution of scores is smaller (height of peak)
How does effect size differ from statistical hypothesis testing?
Unlike statistical hypothesis testing, effect size is a standardized measure based on distributions of scores rather than distributions of means
* Rather than om = o/√N, effect sizes are based only on the variability in the distribution of scores and do not depend on sample size
Since effect sizes are not dependent on sample size, what does this allow us to do?
This means we can compare the effect sizes of different studies with each other, even when the studies have different sample sizes
When we conduct a z-test, the effect size is typically
Cohen’s D: a measure of effect size that expresses the difference between two means in terms of SD
* AKA, Cohen’s d is the standardized difference between two means
Formula for Cohen’s d for a z statistic:
d = (M - u)/o
- Similar to z statistic (om -> o, um -> u)
With the results, we can determine (from Cohen’s 3 guidelines)…
Small, Medium, Large Effects
- Small effects: 0.2 | 85% overlap
- Medium effects: 0.5 | 67% overlap
- Large effects: 0.8 | 53% overlap
Does an effect need to be large to be meaningful?
Just because a statistically significant difference is small, that does not necessarily suggest no meaning; interpreting the meaningfulness of the effect sizes depends on the context
Meta-analysis:
Meta-analysis
- a study that involves the calculation of a mean effect size from the individual effect sizes of more than one study