Critiquing research findings Flashcards

(32 cards)

1
Q

descriptive statistics

A

describe/synthesize data about the sample and the study variables

  • frequency distribution
  • measures of central tendency (mode/mean/median)
  • measures of dispersion (range/variance/SD)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

inferential statistics

A

make inferences about population based on sample data

test hypotheses

answer research questions

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

parametric statistics

A

a class of statistical tests that involve assumptions about the distribution of the variables and the estimation of a parameter

data are NORMALLY DISTRIBUTED

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

nonparametric statistics

A

a class of statistical tests that do NOT involve stringent assumptions about the distribution of critical variables

data are NOT NORMALLY DISTRIBUTED

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

null hypothesis

A

rejected if relationship is statistically significant (p < 0.05)

accepted if relationship is NOT statistically significant (p = 0.05 or greater)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

p-value

A

probability of rejecting the null hypothesis when the null is actually true

typically, p<0.05 is real effect

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

confidence interval (CI)

A

range of values within which a population parameter is estimated to lie, at a specified probability (eg 95% CI)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

confidence limit

A

upper/lower boundary of a CI

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

correlation statistics

A
  • indicate direction and magnitude of relationship between 2 variables
  • used with ordinal/interval/ratio measures
  • can be shown graphically (scatter plot)
  • correlation coefficient can be computed
  • with multiple variables, a correlation matrix can be displayed
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

bivariate correlation

A

2 variables

  • Pearson’s r, a parametric test (lowercase “r” indicates a correlation b/w 2 variables)
  • tests that the relationship b/w 2 variables is not zero
  • used when measures are on an interval/ratio scale (continuous level data)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

strengths of relationships

A

weak: 0.00~0.30 (+ or -)
moderate: 0.30~0.50 (+ or -)
strong: >0.50 (+ or -)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

nonparametric alternatives to bivariate correlation analysis

A
  • Spearman’s rank-order correlation coefficient: measures association b/w ordinal-level variables
  • Kendall’s tau: measures association b/w ordinal-level variables
  • Cramer’s V: measures association b/w nominal-level variables
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

factor analysis

A
  • examines interrelationships among large #s of variables to reduce them to a smaller set of variables
  • IDs clusters of variables that are most closely linked together
  • typically used to assist with validity of a new measurement method or scale
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

simple linear regression

A
  • provides a means to estimate the value of a dependent (outcome) variable based on the value of an independent variable (predictor)
  • outcome variable is continuous (interval/ratio-level data)
  • predictor variables are continuous or dichotomous (dummy variables)
  • change in Y given a one unit change in X
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

multiple linear regression

A
  • predicts a dependent variable based on 2+ independent variables (predictor)
  • dependent variable is continuous (interval/ratio-level data)
  • predictor variables are continuous or dichotomous (dummy variables)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

simultaneous multiple regression

A

enters all predictor variables into regression equation at the same time

17
Q

stepwise multiple regression

A

enters predictors in a series of empirically determined steps, in the order that produces the greatest increment to R^2

18
Q

hierarchical multiple regression

A

enters predictors into equation in a series of steps, controlled by researcher

19
Q

multiple correlation coefficient (R)

A
  • is the correlation index for a dependent variable and 2+ independent variables
  • does not have negative values: shows strength of relationship, but not direction
  • can be squared (R^2) to estimate proportion of variability in dependent variable accounted for by independent variables
  • can’t be less than highest bivariate correlation b/w the dependent v and an independent v
20
Q

power analysis

A
  • method of reducing risk of Type II errors and estimating their occurrence
  • if power = .80, risk of Type II error is 20%
  • estimates how large a sample is needed to reliably test hypotheses

4 components:

  • significance criterion (α)
  • sample size (N)
  • population effect size (γ): magnitude of relationship b/w research variables
  • power: probability of obtaining a significant result (1 - β)
  • generally need 20-30 subjects per independent v
  • <10 subjects per independent variable leads to serious error
21
Q

odds

A

based on probabilities

probably of occurrence/probability of nonoccurrence

22
Q

odds ratio (OR)

A

ratio of odds for the treated vs. untreated group, with the odds reflecting the proportion of people with the adverse outcome relative to those without it

23
Q

dichotomous outcome variable

A

mean of dependent (outcome) variable will be between 0-1

OR = 1 probability of event is same for both groups
OR >1 probability is higher for subjects exposed
OR <1 probability of event is lower among subjects exposed

(slides have more interpretations)

24
Q

commonly used parametric stats

A

student’s t-test

analysis of variance (ANOVA)

25
commonly used nonparametric stats
Mann-Whitney U test: analogous to parametric t-test for independent groups for group comparisons (2 groups) Kruskal Wallis test: analogous to parametric ANOVA (more than 2 groups)
26
t-test
testing for significant differences b/w group means of 2 samples - for independent groups (b/w subjects): examines differences b/w 2 independent groups - for dependent groups (paired/within subjects): matched or paired groups - comparison of pretest and posttest measurements
27
ANOVA
- compares data b/w 2+ groups/conditions to investigate the presence of differences b/w those groups - the outcome or dependent variable must be continuous (interval/ratio) - independent (predictor) variable is nominal - F test: typically accompanied by an ANOVA - doesn't say which groups differ from one another, just that they're different
28
chi-square test
nonparametric test compares differences in proportions of nominal level variables compares actual/observed frequency with expected frequency
29
Cohen's effect sizes for the magnitude of tx effect
``` small = 0.2 moderate = 0.5 large = 0.8 ```
30
power of a test
probability of detecting a difference or relationship if such a difference or relationship really exists anything that dec probability of a Type II error inc power, vice versa a more powerful test is one that's likely to reject H0
31
statistical significance
these results say nothing about clinical importance or meaningful significance of results researcher must always determine if statistically significant results are substantively meaningful
32
post-hoc analyses
performed in studies w/ more than 2 groups when analysis indicates that groups are significantly different but doesn't indicate which groups are different developed specifically to determine location of group difference after ANOVA is performed - Scheffe - Bonferroni - Least Significant Difference (LSD) Test - Tukey's Honestly Significant Difference (HSD) - Student Newman-Keuls - Tukey's Wholly Significant Difference (WSD)