T-tests Flashcards

(43 cards)

1
Q

The Principle of Hypothesis Testing

A

Start with a default assumption, the null, and ask whether the evidence from the sample is strong enough to reject it

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Null Hypothesis

A

H0: μ = 50, for example

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Alternative Hypothesis

A

H1: μ ≠ 50, for example

reject or retain H0

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

α (level of significance)

A

the amount of error the experimenter is willing to tolerate.

when estimating a parameter from a sample statistic it is not possible to eliminate error, only to quantify the likelihood of error

typically set at .05 if not otherwise stated

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Decision Rule

A

Reject H0 if the probability of obtaining a sample mean as deviant or more deviant than the one observed, when H0 is true, is less than or equal to α

convert M to zo first, z = M - μ0 / σM

Decision rule: Reject H0 if |Z| ≥ Zc

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Decision Rule and Deciding α

A

Considering the decision rule changes what we consider to reject or retain by the mean, α becomes a criterion for rejecting or retaining H0

you might think why not choose .01 to minimise the most errors, well, there are multiple types of errors

i.e. we want to strike a balance between type 1 and 2 errors

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Type 1 Error

A

Rejecting H0 when H0 is true

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Correct Acceptance

A

Retain H0 when H0 is true

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Correct Rejection

A

Reject H0 when H0 is false

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Type II Error

A

Retain H0 when H0 is false

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Which Error is regarded as worse?

A

Type 1 errors, so the emphasis is on controlling for this. People almost always set α at .05 or less

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Degrees of Freedom

A

the number of independent observations in a set of data

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Principle of T-tests

A

When we don’t know the population mean (μ)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Note on sM (instead of σM)

A

there is no longer a normal distribution using the sample

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Properties of t distribution

A

as df increases, t distribution becomes closer in shape to a normal distribution

  • for lower values of df, t distribution has narrowed peak and fatter tails than normal distribution, and therefore has a larger standard deviation
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Confidence Intervals for μ (σ unknown)

A

μ upper = M + (tc x sM)

μ lower = M - (tc x sM)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

T-test formula for a single mean

A

√ Σ(X-M)^2 / n (n-1)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

Paired Samples Design for Two Means

A

same participant on multiple occasions, measuring one participants change over time (within groups)

test of dependent means

19
Q

Independent Samples Design for Two Means

A

cases are independent from each other (between-groups)

independent means

separate people placed in different groups

20
Q

Difference Score Method

A

X1 - X2 = XD

XD - MD

21
Q

Paired Samples Equation

A

sMD = √Σ(XD-MD)^2 / n(n-1)

22
Q

Confidence Interval for Paired samples

A

uD upper = MD + (tc x sMD)

uD lower = MD + (tc x sMD)

we can be 95% that population mean … is between … and … under … condition than … condition

23
Q

P-value

A

the probability of obtaining a t statistic that is deviant to the one obtained, even when the null is true (under h0)

probability that the result occurred by chance when there was no true effect

p<0.05 could be considered statistically significant

24
Q

Natural Pairs

A

-The researcher does not randomly allocate participants to one group or another; the pairing occurs ‘naturally’ prior to the study

  • Although the pairs are made up of different participants the scores of each pair are likely to be correlated → dependent means variable
  • Example: researcher wants to compare marital satisfaction of husbands and wives
25
Matched Pairs
- The researcher has control over the ways the pairs are formed so that participants are matched on some variable - Although the pairs are made of different participants the scores of each pair are correlated → dependent means analysis - Example: examine effects of stress on food intake in rats, participants divided into pairs matched on weight, and within pairs are randomly allocated to groups (stressed and non-stressed group), two groups won't have a third variable for weight
26
Standard Error Formula when finding difference between means (M1-M2)
σ M1-M2 = √ σ1^2/n1 + σ 2^2/n2
27
Formula for Independent Samples Mean
S M1-M2 =√Σ(X1-M1)^2 +Σ(X2-M2)^2 / (n1-1) +(n2-1) x (1/n1 + 1/n2)
28
Independent Scores Method
X1-M1 + X2-M2
29
Confidence Intervals for (μ1-μ2)
(μ1-μ2) upper = (M1-M2) + (tc X S m1-m2) (μ1-μ2) lower = (M1-M2) - (tc X S m1-m2)
30
Finding t for Independent Means
t = (M1-M2) / Sm1-m2
31
Assumptions for Independent Means T-tests
1. Normal distribution of scores in each population 2. Population variances are equal 3. Observations are independent, within and between groups (one person's productivity while listening to music does not influence another person's productivity while listening to music)
32
Changes in Outcomes if Independent t-test assumptions are met or not
- When all 3 met, Type 1 error rate = nominal rate (a) (the rate expected by the experimenter) - When assumption are violated, actual distribution of t values does not follow theoretical t distribution - As a result, actual type 1 error rate does not equal a
33
Robustness
a test is said to be robust against violation of an assumption if the actual type 1 error rate is close to α even when that assumption is violated - evaluated using monte carlo simulations
34
What is a t-test robust against?
violation of normality assumption - central limit theorem - sampling distribution will always be normal if the sample size is large enough violation of assumptions of equal variances provided n1=n2 - when unequal outcomes depends on which group has larger variance - When the larger sample is associated with larger variance, the t-test is conservative (think you have a big standard error when you actually don’t) (less type 1 errors, but higher risk of type 2 errors) - When the larger sample is associated with the smaller variance, the t-test is liberal (more type 1 errors, less risk of type errors) (type 1 error rate will go above alpha) (type 1 errors are more important)
35
What is the t-test not robust against?
Violation of independence of observations (when the data points are unrelated to each other - groups tested together, such as comparing 2 classrooms - Will not be a good experiment, naturalistic, quasi-experiment - Here the behaviour of one participant may influence the behaviour of other participants - May be a very good or bad student that may have an influence on all the scores in that condition - Classes may not be equal in their academic ability etc
36
EV's
extraneous variables - any variable that can effect the dependent variables of your study intrinsic - participant variables - temporary EV's (mood, motivation, etc) - relatively stable EV's (intelligence, ability, personality) other EV's relate to the environment - testing conditions, measurement error
37
Are EV's relative?
one researchers EVs may be another researchers IVs
38
Controlling for EV's
- Internal validity - an experiment is internally valid if there are no other explanations for changes in the DV, other than the manipulation of the IV EV’s are influences on the DV not due to the IV, hence threaten internal validity - Need to - Make sure the only systematic difference between conditions is the IV - Reduce amount of error in data due to EVs i.e through group variability
39
Random Allocation for Independent Samples
Controlling for EV's related to participants - Random allocation is used to convert possible systematic errors into random errors - Provides some control over all potential participant EVs, random allocation deals with them whether we know about it or not - Does not equate groups; rather random allocation distributions EVs impartially between groups Rather than random sampling, which increase external validity, as we get a diverse and impartial population base
40
Controlling EV's for Paired Samples
- Measuring same participants (repeated measure) or matched pairs produces a paired sample design - Ensures that distribution of scores on EVs related to participants are held constant from one condition to the next (individual differences are not applicable, because they are the same for each condition, because it is the same person) - This can reduce the standard error for the difference between conditions - Manages the variability of participant performance by measuring the difference between conditions for each individual participant
41
Issue of Rank-Order Effects for EV's in Paired Samples
Rank order effects: conditions presented earlier may be responded to differently than conditions presented later - Positive rank order effects: practice, learning, reduction in symptom - Negative rank order effects: fatigue, boredom - These are problematic if conditions are confounded with order (everyone does condition X first and Y second - Solved with Counterbalancing - E.g if there are two conditions A and B, Half do A first B second, Half do B first A second - Random permutations: each participant does a random order of conditions
42
Issue of Carry Over Effects in EV's for Paired Samples
- Effect of one condition carry to the next - Does not necessarily have to do with a specific order, only one condition - Not solved by counterbalancing
43
Comparing Strength of Paired VS Independent
In paired samples, difference scores are normally lower in variance, correlation between scores in condition 1 and condition 2 should be positive, less random variability than independent sample - Standard error is smaller than when samples are independent - Therefore, we get a larger t - Thus, we are more likely to be rejected with paired samples design than independent - Dependent t test is a more powerful test of accepting the null than independent i.e more likely to lead to rejection of a false H0