Week 4 - Lecture 4 Flashcards

1
Q

what is statistical power?

A

The probability that a test will detect an effect if one exists (ie. correctly rejecting a false null hypothesis)
- formula: power = 1 - β (Type II error rate)
- power increases the chances of detecting a true effect

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

what does the null hypothesis (H0) state in a study?

A

That there is no effect or no difference

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

what does the alternative hypothesis (H1) state?

A

That there is an effect or a difference exists

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is a Type I error?

A

Saying there is an effect when there isn’t one (false positive)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is the symbol used to represent a Type I error?

A

Alpha (α)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What is the common value set for alpha (α) in research?

A

0.05 (or 5% chance of making a Type I error)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Give an example of a Type I error

A

A pregnancy test says someone **is pregnant **when they are **not **

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is a Type II error?

A

Saying there is no effect when there actually is one (false negative)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What is the symbol for a Type II error?

A

Beta (β)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Give an example of a Type II error

A

A pregancy test says someone is not pregnant when they actually **are **

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What does **power **refer to in significance testing?

A

The ability to correctly detect an effect when it exists (1 - β).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What happens to power when beta increases?

A

power decreases (you’re more likely to miss real effects)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

In terms of errors, why do we not always reduce alpha to 0.01?

A

Because it would increase beta and make us more likely to miss real effects (Type II errors)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What does it mean if power = 0.80?

A

There is an 80% chance of correctly finding a real effect if it exists

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

complete the table:

A

refer to photo

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Does statistical significance always mean the result is practically important?

A

No. A result can be statistically significant but still be** too small to matter in real life**

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

How does a large sample size affect statistical significance?

A

It can make tiny effects appear **statistically significant **

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

What can happen with a small sample size?

A

It might miss real effects because there’s not enough data

19
Q

What is an example of a study with a statistically significant but practically unimportant result?

A

Facebook’s 2014 study - effect size was d = 0.02, which is very small.

20
Q

What happens to power if you lower alpha (eg. from 0.05 to 0.01)?

A

Power decreases (but you reduce the chance of a Type I error).

21
Q

What happens to power if you increase alpha (eg. from 0.05 to 0.10)?

A

power increases (but so does the chance of a Type I error)

22
Q

What does “effect size” tell us?

A

How big the difference or relationship is, measured in standard deviation units

23
Q

What are Cohen’s effect size benchmarks for small, medium and large effects?

A
  • small = 0.2
  • medium = 0.5
  • large = 0.8
24
Q

How does sample size affect power?

A

larger sample size gives more power by reducing standard error

25
What is the effect of increasing sample size from N = 25 to N = 100 on significance?
It can change a non-significant result (Z=1.5) to a significant one (Z=3)
26
What happens to power when variance in the population is low?
power increases because it becomes easier to detect real differences?
27
How can you reduce variance in your sample?
By using a homogeneous (similar) group of participants
28
What does effect size (d) measure?
The **difference between group means**, expressed in standard deviation (SD) units
29
What does statistical effect (δ or delta) represent?
How many standard errors apart the populations are
30
What's the formula for δ in a one-sample t-test?
refer to photo
31
What’s the formula for δ in a two-sample t-test (equal group sizes)?
refer to photo
32
What is the formula to calculate effect size (d)?
refer to photo
33
What do we use δ for once it’s calculated?
We use it to look up power in a power table, given a chosen α level.
34
If d = 0.333 and N = 25, what is δ and power approximately?
- δ = 1.665 - Power ≈ 0.36 (not sufficient)
35
If d increases to 0.667, what happens to power (with same N)?
δ increases to 3.335 and power ≈ 0.91 → power improves
36
What is the formula to find sample size N for one-sample tests?
refer to photo
37
What value of δ is commonly used to achieve 80% power?
δ = 2.8 (when α = 0.05)
38
If d = 0.333 and δ = 2.8, what is the required N?
refer to photo
39
Why must you always round up when calculating N?
You can't test a fraction of a person, and rounding down gives less power than needed
40
In two-sample tests, what does lowercase n represent?
The number of people in each group
41
What is the formula for n in a two-sample t-test?
refer to photo
42
If d = 0.5 and δ = 2.8, how many people are needed in each group? How many total participants are needed in that example?
63 per group Total N = 63 x 2 = 126 participants
43