Research Basics pt. 2 Flashcards

(56 cards)

1
Q

Parameter

A

-descriptive value for a population

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Statistic

A

-descriptive value for a sample

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Mean

A

-average
-most commonly used
-only used with interval/ratio
-influence by outliers
-toward the tail oppositte of mode

-μ, x

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

when shouldn’t you report the mean?

A

if you have outliers/extreme scores, the mean will be pulled towards the extremes (towards the tail) and will not provide a central value.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Variance

A

-SD^2 or (distance from mean)^2/ n-1

-σ^2

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Standard Deviation

A

-distance between a score and mean
- Square root (distance from mean)^2/ n-1

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Frequency Distribution

A

-organized picture of an entire set of scores

-histogram, smooth curve, stem and leaf

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Smooth Curve

A
  • emphasizes the fact that the distribution is NOT showing the exact frewuency for each category
    -want it to be symmetrical (normal curve, mean and median are equal)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

1 SD in a normal distribution

A

68.26% (34.13%)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

2 SD in a normal distribution

A

95.44% (13.59%)

68-95-99.7 Rule

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

3 SD in a normal distribution

A

99.72% (0.13%)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Histogram

A

-shows all the frequencies of the distribution

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

positive vs. negative skew

A

-non symmetrical distribution
-named for tail

Positive: scores pile up at low values, tail point to high values

Negative: scores pile ip at high values with tail at low

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Kurtosis

A

-peakedness of tthe distrubution

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Leptokurtic

A

-skyscraper
-higher and thinner peak
-low variability
-easier to get significance

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Platykurtic

A

-hill
-lower peak
-higher variability
-harder to get significance

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

Stem-And-Leaf Display

A
  • preserves the original data values
  • It’s especially useful for small to moderately sized data sets.
    -each score devided into a stem (first digit) or leaf (last digit)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

central tendency measures

A

describes the center of the distribution and represents the entire distribution of scores as a single number (mode, median, mean)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

Mode

A

-most frequent
-used in all data
-located on one side near peak, other farthest from mean
-bimodal, multimodal

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

Median

A

-middle: 50% of the scores in the distribution have values that are equal or less than the median
-used for ordinal, intterval, or ratio
-unnaffected by outliers
-can’t show sig dif
-between mean and mode

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

Variabiltiy

A

-how spread out the data is
-descriptive (how spread out) and inferential stats (how accurate to pop)
-meaured by range or SD

less variability –> better representation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

Range

A

-total distance

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

SD in Normal Distribution

A

-70% of scores 1 SD of mean (35+/-)
-95% of scores 2 SD of mean
-99% of scores 3 SD of mean

standardized, mean is 0

24
Q

Z Score

A

-where a score is located relative to other scores
-# of SD above or below mean
-descriptive (where in curve) and inferential stats (reference to population)

z= score-mean/SD

25
Inferential Statistics
-infer things about the population based on sample
26
Probability
-proportion under the curve -z score creates % as body or tail
27
Critial Limit Theorem
-30 sample with be closly related to real pop
28
T-Test
-compare 2 groups -used fo smaller samples -flater curve than normal distibution -1 tail only considers 0.05 in one tail= higher chance of significance -2 tail considers 0.05 in both (0.025 in each)= lower chance of significance
29
F-Distribution
-ANOVA -more than 2 or factorial research design
30
Chi-Square Distribution
-comparing proporttions of people in diff groups -comparres observed frequencies to expected
31
Standard Error of Mean (SEM)
-value that describes the diff between the sample mean and true pop mean -always smaller than SD -smaller=less sampling error -sample SD/Square root (n)
32
Point Estimate
-mean of sample, estimates pop -boarder of box-plot
33
Interval Estimate (CI)
-confidence interval -range of sample that can include the real pop mean -span of box plot
34
Box-Whisker Diagram (boxplot)
-Whiskers: range of scores -Box: median (line), upper and lower quatile (25 and 75%)
35
Bar Graph
-nominal or ordinal -similar to hisogram with space
36
Error Bars Charts
-bar shoes mean score Can show -CI, SD, or Stardard error of mean
37
Scatterplots
-correlation -can be grouped (R is important)
38
Parametric Statistics
Analyzes quantitative data -t test, anova, regression -must meet assumptions -based on distributions so must be normalized
39
Non-Parametric Statistics
Analyze qualitative data -spearman, mann-whitney u (independent t test, takes mean rankings), friedman’s ANOVA, wilcoxson (independent t test, takes mean rankings) -violates the assumptions or have nominal/ordinal data
40
Linear Regression
-show relationships -make predictions
41
Parametric Assumptions of T-Test
I/R Data Normality Homogenity of Variance Free of Extreme Outliers Independence of Observations
42
Normality
-concern with smaller studies <30 -check skewednwss ( 2 is a problem) -check histograms Non-Parametric -Shapiro-Wilk Test: >0.5 is not significant
43
Homogeneity of Variance
Differences in variance should be equal Non-Parametric: -Levene’s test: want it to be not significant >0.5
44
Free of Influential Outliers
Regression: cook’s distance (>1 is bad)
45
Independence of Observations
-scores must not follow a pattern over time -scores from one participant cant influence another persons scores Non-pArametrics: Durbin Watson
46
Regression Assumptions
Linearity Homoscedasticity Outlier testing in regression
47
Homoscedasticity
-relationship statistics -seen in a scatterplot’s residual score -variance must be the same at all levels -how close are all points to the line -r^2 -heteroscedasticity is opposite
48
Linearity
-data points arranged in a linear pattern -seen in a scatterplot
49
Residual Score
-distance of score from regression line on y axis -ouliers are large
50
Standardized Residual
-distance from line in terms of SD - negative= under the line -positive = over line -0= on line
51
Solutions to Violated Assumptions
Trim the data Windoring: substitute outlier with highest score Transform the data: take the log of the data Bootstrapping is SPSS: Non parametric Data
52
Critical Region
-in the tails -outcomes unlikely caused by chance
53
How To Increase Power
-increase effect size -decrease variability -increase sample size -increase alpha -use a 1 tail test
54
Independent T-Test
-compares 2 means of independent data -different groups -1 IV and 1 DV -Man Whitney U
55
Repeated Measures T-Test (Dependent)
-compaires matched pairs -same participant twice -more likely to be significant -wilcoxon signed ranks -does not need HOV
56
Bonferroni Correction
-limits alpha inflation when testing the same data set multiple times and makinf a type 1 error - divides alpha by number of tests run