Statistics exam 4 Bayes Flashcards

1
Q

What is the bayes factor?
What does BF10 and BF01 mean?

A

The ratio of two competing models, represented by their evidence. It’s the predictive updating factor

BF10: x times more likely under H1 than H0
BF01: x times more likely under H0 than H1

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What does BF = 1 mean?

A

Both models predicted data equally well

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is the beta distribution and when do we use it?

A

It’s a type of probability distribution for binomial variables

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What determines the shape of the beta distribution?

A

A (successes) and B (fails)

If a = b: centered bell shape
If a < b: shape left centered
If a > b: shape right centered
If a and b <1: mass close to 0 and 1
If a = b = infinity: spike
If a = b = 1: uniform (uninformative)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is parsimony?

A

Specific models are rewarded more when predicting well than non-specific competitors

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What is transitivity?

A

BF (BA) = 2
BF (AS) = 2
So: BF (BS) = 2 * 2 = 4

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is bayes theorem for posterior and prior beliefs?

A

Posterior = prior * predictive updating factor

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is the difference between likelihood and marginal likelihood?

A

Likelihood: likelihood of the data for all theta’s. This creates a shaped distribution. The area doesn’t sum to 1

Marginal likelihood: average across all likelihoods, weighted by density at each point

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What does the following mean:
Likelihood > marginal likelihood
Likelihood < marginal likelihood

A

L > ML = values of theta that predicted data better than average
L < ML = values of theta that predicted data worse than average

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What happens to values of theta that were 0 in the prior distribution after updating? Why is a spike not a good prior model?

A

Values of theta that were 0, can’t be updated, since multiplying by 0 always ends with 0
A spike is blind for updating. The spike has infinitely large value, so multiplying by even the smallest number, will still leave it at infinity.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is truncation?

A

Some values of theta are assigned 0 density, which makes it a one-sided model

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

From what model do you typically start?

A

Uninformative model. The less informed the prior, the more the data can speak for itself

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

How do you update the a and b in the beta distribution?

A

a = a + number of successes
b = b + number of fails

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

How can you estimate a proportion from the posterior? Name two ways

A
  1. Take a median or mean
  2. Credible interval: take middle 95%
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

How do you interpret a credible interval?

A

..% probability the true value of theta is between these borders

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What happens to the prior beliefs when we think both hypotheses are equally likely?

A

The prior belief value will be equal to 1

17
Q

What is the Savage Dickey Density ratio and what is it used for? What does it say?

A

It’s used in bayesian hypothesis testing. It’s the ratio of prior/posterior

If prior > posterior –> evidence for H1

18
Q

What is the test parameter of a bayesian test?

A

The bayes factor

19
Q

What does =/ , > and < mean for the shape of the beta distribution in testing?

A

=/ : uniform
< or > : truncated

20
Q

What is the difference in winnings between one-sided and two-sided hypotheses?

A

Two-sided spreads bets more and therefore receives less winnings –> lower marginal likelihood –> lower BF

One-sided has more winnings if it’s correct. Higher marginal likelihood –> higher BF

Parsimony!

21
Q

What is sequential analysis?

A

Updating beliefs and seeing evolution of the BF. There is usually more support after n>30

22
Q

What is important for choosing a prior?

A
  • Informed by previous knowledge
  • One/two sided: do you want to know a difference or specific difference
  • Same domain
23
Q

What are the domains for proportion, correlation and difference in means?

A

Proportion : 0-1
Correlation: -1 to 1
Means: - infinity to infinity

24
Q

What distribution does bayesian correlation use?

A

Stretched beta distribution with fitting domain

25
Q

What happens to the credible interval and the Bayes factor when adding more participants?

A

CI gets more narrow
BF gets higher

26
Q

What test statistic does the bayesian t-test use?

A

Cohens d (delta sign): standardized difference between groups

27
Q

What distribution does the bayesian t-test use? What determines its shape?

A

Cauchy distribution. Shape is determined by the width

There is no uniform distribution possible because the range is infinitely large

28
Q

What does the width of the Cauchy distribution indicate?

A

It indicates the area where 50% of the values are

E.g w = 0,7 –> 50% between -0,7 and 0,7

29
Q

What is a robustness check?

A

Sensitivity analysis for the T-test. It explores what would have happened with a different prior width

BF is pretty stable for many prior widths, except strong prior settings

30
Q

What are two initiatives that emerged from the reproducibility crisis?

A
  1. Preregistration
  2. Open science
31
Q

What are the 6 steps for Bayesian testing?

A
  1. assumptions
  2. hypotheses
  3. set prior distribution
  4. compute likelihood
  5. bayes factor
  6. conclusion
32
Q

What are the 6 steps for frequentist testing?

A
  1. set alpha
  2. assumptions
  3. hypotheses
  4. test statistic
  5. p value
  6. conclusion