ARMS Week 1 Flashcards

(33 cards)

1
Q

Frequentist approach vs Bayesian framework

A

Frequentist:
- Test how well data fit H0
- p-values, confidence intervals, effect sizes, power analysis

Bayesian:
- Probability of hypothesis given the observed data -> prior information
- BFs, prior, posterior, credible intervals

Both can be used for estimation and hypothesis testing.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Frequentist estimation

A
  • Emprical research uses collected data to learn from
  • likelihood function
  • probability of an event = frequency it occurs
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Bayesian estimation

A
  • prior information about u
  • prior is updated and provides posterior distribution of u
  • conditional probabilities -> P(A given B)
    Pro: accumulating knowlegde
    Con: results depend on choice prior
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Posterior mean or mode

A

The mean or mode of the posterior distribution. Should add up to 1, just as the prior model probabilities.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Posterior SD

A

Standard deviation of posterior distribution (comparable to frequentist standard error)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Posterior 95% credible interval

A

Providing bounds of the part of posterior with 95% of posterior mass

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Posterior Model Probability (PMP)

A

The probability of the hypothesis after observing the data
Depends on two criteria:
1. How sensible it is, based on the prior
2. How well it fits the new data

  • PMP are relative probabilities. They are updates of prior probabilities with the BF.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Bayesian testing is comparative. What does this mean?

A

Hypotheses are tested against one another, not in isolation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What does the Bayes factor say?

A

How much support there is for H1 compared to H0.
BF10 = 10 -> support for H1 is 10 times stronger than for H0.
BF10 = 1 -> support for H1 is as strong as for H0.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Probability theory: frequentist vs bayesian

A

Frequentist:
- probability is the relative frequency of events

Bayesian:
- probability is the degree of belief

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

95% Confidence interval vs 95% credible interval

A

Confidence interval 95% (frequentist)
- if we were to repeat this experiment many times and calculate a CI each time, 95% of the intervals will include the true parameter value.

Credible interval 95% (Bayesian)
- there is a 95% probability that the true value is in the credible interval
- a zero present = not significant

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Linear regression

A

Lineair association between a dependent and independent variable

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Residual

A

Difference between value and the line in the plot (we can’t explain them).
- we want to minimize residuals

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Multiple Lineair regression (MLR)

A

Lineair regression with multiple predictors (independent variables)

Assumptions:
- Dependent variable is continuous (interval/ratio)
- Independent variables are continuous or dichotomous (nominal with two categories)
- Linearity of relations (the L in MLR) -> checked with scatterplots
- No outliers

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Dummy variables

A

Has value 0 and 1. Is used in MLR to code data suitable for this approach (interval/ratio).
Number of dummy variables is equal to the groups minus 1

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

R^2 vs adjusted R^2

A

R^2 -> proportion of explained variance in the sample
Adjusted R^2 -> proportion of explained variance in the population

17
Q

R

A

Correlation coefficient, explains how much the variables correlate with one another

18
Q

B value

A

Unstandardized effect

19
Q

B0

20
Q

Hierarchical MLR

A

Comparing two nested models (multiple formulas) -> one model is a smaller version of the other

21
Q

R^2 changed

A

How much more variance does the second model explain compared to the first?

22
Q

Prior

A

Existing knowlegde before looking at own data

23
Q

Prior model probabilities

A

How likely is each hypothesis before seeing the data.
- add up to 1 -> relative probabilities divided over the hypotheses of interest
For example: H1 = 0,2 and H2 = 0,8 ; or H1 = H2 = H3 = 0,33

24
Q

Standardized residuals

A

Check wether there are outliers in the Y-space.
Rule of thumb -> values must be between -3,3 and +3,3

25
Cook’s Distance
With this you can check whether there are outliers within the XY-space. Rule of thumb -> must be lower than 1
26
Multicollinearity
Indicates whether the relationship between two or more independent variables is too strong. Including overly related variables in your model has three consequences: 1. B is unreliable 2. Limits magnitude of R (correlation between Y and Y^) 3. Importance of individual independent variables can hardly be determined (if at all) Possible solutions: - remove variables - combine variables in scales
27
Variance Inflation Factor (VIF)
Used to determine whether multicollinearity is an issue. Rule of thumb -> VIF > 5 = potential problem, VIF > 10 = problem.
28
Homoscedasticy
Spread of residuals must be approximately the same across all values for y. When not: heteroscedasticy
29
B (Beta)
Standardized coefficients, can be used to determine the most important predictor of a model. The independent variables with the largest beta is the most important predictor. -> - or + doesn’t matter.
30
Construct validity
Extent to which a conceptual variable is accurately measured or manipulated
31
Internal validity
Extent to which the research method can eliminate alternative explanations for an effect/relationship (trying to find causal relationship)
32
External validity
Extent to which the research results generalize to besides those in the original study
33
Statistical validity
Extent to which te results of a statistical analysis are accurately measured and well founded (checking assumptions and report significance)