Exam 2 Flashcards

(87 cards)

1
Q

What are dependent variables and what are examples?

A

The outcome variable measured

May be changed or influenced by manipulation of the independent variable

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What are conceptual defined dependent variables?

A

Conceptual variable - the general idea of what needs to be measured in a study: i.e. strength

Must be defined in specific measurable terms for use in the study (once defined, becomes the operational variable)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What are the levels of precision of measurement and what are examples?

A

The level of measurement/precision of the dependent variable is one factor that determines the choice of statistical tests that can be used to analyze the data

  • Nominal (Count Data)
  • Ordinal (Rank Order)
  • Interval
  • Ratio
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is reliability and why is it important?

A

Reliability: An instrument or test is said to be reliable if it yields consistent results when repeated measurements are taken

3 Forms of Reliability:

  • Intra-rater reliability
  • Inter-rater reliability
  • Parallel-forms reliability

All three forms require comparisons of two or more measures taken from the same set of subjects

Requires a statistical measure of this comparison – correlation

Results of these statistical tests are 0 to 1, with 1 being a perfect correlation and 0 being no correlation

For a reliable comparison the association must be as close to 1 as possible and at least > 0.8

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is validity and why is it important?

A

Validity - the measure or instrument (measuring tool) is described as being valid when it measures what it is supposed to measure

A measure/instrument cannot be considered universally valid.

  • Validity is relative to the purpose of testing
  • Validity is relative to the subjects tested

Validity is a matter of degree; instruments or tests are described by how valid they are, not whether they are valid or not

Validity assessments can be made by judgment (judgmental validity) or by measurements (empirical validity)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Judgmental Validity

A

Based upon professional judgment of the appropriateness of a measurement or instrument

2 principle forms:

  • Face validity
  • Content validity
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What are Sensitivity, Specificity and Predictive Values?

A

Clinical research often investigates the statistical relationship between symptoms (or test results) and the presence of disease

There are 4 measures of this relationship:

  • Sensitivity
  • Specificity
  • Positive Predictive Values
  • Negative Predictive Values
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What are extraneous/confounding variables and what are examples?

A

Confounding or Extraneous Variables: Biasing variables – produce differences between groups other than the independent variables

These variables interfere with assessment of the effects of the independent variable because they, in addition to the independent variable, potentially affect the dependent variable

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Analysis of the Dependent Variable

A

A single research study may have many dependent variables

Most analyses only consider one dependent variable at a time

Univariate Analyses- each dependent variable analysis is considered a separate study for the purposes of statistical analysis

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What are operationally defined dependent variables and what are examples?

A

The specific, measurable form of the conceptual variable

The same conceptual variable could be measured many different ways; defined in each study

i.e: to define strength = Maximum amount of weight (in kilograms) that could be lifted 1 time (1 rep max)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Nominal Level of Precision

A

Score for each subject is placed into one of two or more, mutually exclusive categories

Most commonly, nominal data are frequencies or counts of the number of subjects or measurements which fall into each category.

Also termed discontinuous data because each category is discrete and there is no continuity between categories

There is also no implied order in these categories

Examples:

  • Color of cars in parking lot
  • Gender of subjects
  • Stutter – yes/no

Can only indicate equal (=) or not equal (≠)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Ordinal Level of Precision

A

Score is a discrete measure or category (discontinuous) but there is a distinct and agreed upon order of these measures, categories

The order is symbolized with the use of numbers, but there is no implication of equal intervals between the numbers

Examples include any ordered scale

Can say = & ≠, and greater than (>) or less than (<)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Interval Level of Precision

A

Numbers in an agreed upon order and there are equal intervals between the numbers

Continuous data ordered in a logical sequence with equal intervals between all of the numbers and intervening numbers have a meaningful value

There is no true beginning value – there is a 0 but it is an arbitrary point

Examples:

  • Temperature
  • Dates
  • Latitude (from +90° to −90° with equator being 0

Interval level can say = or ≠; < or >; and how much higher (+) or how much lower (−)

(Interval does not have a true 0 value; Ratio does)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Ratio Level of Precision

A

Numbers in an agreed upon order and there are equal intervals between the numbers

Continuous data ordered in a logical sequence with equal intervals between all of the numbers and intervening numbers have a meaningful value

There is a true beginning value – a true 0 value

Ratio level – can do everything with interval but also can us terms like 2X or double (×) or half (÷) when comparing values

(Interval does not have a true 0 value; Ratio does)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Mathematical Comparison of Precision Levels

A

Nominal level – can only indicate equal (=) or not equal (≠)

Ordinal level – can say not only = & ≠ but also greater than (>) or less than (< or >; and how much higher (+) or how much lower (−)

Interval level can say = or ≠; < or >; and how much higher (+) or how much lower (−)

Ratio level – can do everything with interval but also can us terms like 2X or double (×) or half (÷) when comparing values

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Statistical Reasoning for the Precision Levels

A

Nominal level – unit of central tendency is the mode and any variability is assessed using range of values

Ordinal level – unit of central tendency is the median and any variability is assessed using range of values

Interval or Ratio (Metric) - unit of central tendency is the mean and any variability is assessed using standard deviation if normally distributed

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

Intra-Rater Reliability

A

Tested by having a group of subjects tested and then re-tested by the same person and/or instrument after an appropriate period of time

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

Inter-Rater Reliability

A

Reliability between measurements taken by two or more investigators or instruments

Different instruments or different investigators

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

Objective Testing

A

Usually has both high intra-rater and inter-rater reliabilities

Example: isokinetic dynamometer to measure muscle force (torque)

Raters are trained in the use and calibration of the instrument

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

Parallel Forms Reliability

A

Specific type of reliability used when the initial testing itself may affect the second testing.

Two separate forms of the exam covering the same material, or testing the same characteristic, are used

Investigators must first do testing to demonstrate that the two forms of the test are, in fact, equivalent

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

Face Validity (Judgmental)

A

Judgment of whether an instrument appears to be valid on the face of it - on superficial inspection, does it appear to measure what it purports to measure?

To accurately assess this requires a good deal of knowledge about the instrument and what it is to measure.

That is why professional judgment is emphasized in the definition of judgmental validity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

Content Validity (Judgmental)

A

Judgment on the appropriateness of its contents.

The overall instrument and parts of the instrument are reviewed by experts to determine that the content of the instrument matches what the instrument is designed to measure

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

Empirical Validity (Criterion-Related Validity)

A

Use of data, or evidence, to see if a measure (operational definition) yields scores that agree with a direct measure of performance

2 forms of Empirical Validity:

  • Predictive
  • Concurrent
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
24
Q

Predictive Validity (Empirical)

A

Measures to what extent the instrument predicts an outcome variable.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
25
Concurrent Validity (Empirical)
Measures the extent to which the measures taken by the instrument correspond to a known Gold Standard Correlation statistics are also used to measure the validity
26
Inter-Relationship between Reliability and Validity
For an instrument or test to be useful it must be both valid and reliable
27
Sensitivity
Formula: Sensitivity = TP/TP+FN Definition: the probability that a symptom is present (or screening test is positive) given that the person does have the disease Patient Relevance\*: “I know my patient has the disease. What is the chance that the test will show that my patient has it?”
28
Specificity
Formula: Specificity = TN/TN+FP Definition: the probability that a symptom is not present (or screening test is negative) given that the person does not have the disease Patient Relevance: “I know my patient doesn’t have the disease. What is the chance that the test will show that my patient doesn’t have it?”
29
Positive Predictive Value
Formula: TP/TP+FP Definition: the probability that a person has the disease given a positive test result Patient Relevance: “I just got a positive test result back on my patient. What is the chance that my patient actually has the disease?”
30
Negative Predictive Value
Formula: TN/TN+FN Definition: the probability that a person does not have the disease given a negative test Patient Relevance: “I just got a negative test result back on my patient. What is the chance that my patient actually doesn’t have the disease?”
31
What is the definition and use of descriptive statistics?
Statistics used to characterize shape, distribution, central tendency & variation (variability) of a population or sample
32
What is the parameter & what is the statistic?
Parameter –is the individual value or data point Statistic – calculated descriptive index
33
What are measures of central tendency?
•Distributions described as the variation pattern around the point of central tendency Measures of central tendency: * Mean – the arithmetic mean which is the sum of all values in the distribution divided by the number of values * Median – the middle value in a distribution of values; half way in number from the lowest to highest numbers * Mode – value in the distribution with the highest frequency – most common number in the distribution
34
What is and what measures are there of score distribution and frequency distribution?
Frequency distribution – rank order of how many times a particular value occurred Can be displayed as percentage value and cumulative percentage Data can be grouped into classes or ranks
35
How can frequency distributions be displayed graphically?
Can be displayed as histogram or frequency polygon
36
Confidence Interval
There will always be some errors in clinical research; confidence intervals attempt to allow for this by giving a +/- range of values. CI 95% means that the researcher is 95% confident that the true value is between a given range. Confidence intervals are defined in the research paper
37
Distribution
Pattern of values of a set of scores with n = total number
38
Histogram
Representation of tabulated frequencies, shown as adjacent rectangles, in discrete intervals, with height equal to the frequency of the observations in the interval
39
Frequency Polygon
Intervals represented as points and level on vertical axis represents the frequency of the observations in the interval
40
Normal Distribution
Variation of the distribution pattern around the point of central tendency: Symmetric distribution of values of around the point of central tendency Mode, median & mean all the same point
41
Skewed Distribution
Asymmetric distribution of values of around the point of central tendency Described as positive or negative skew based upon where majority of values are compared to the mode Use “tail” of distribution to determine skew Median & mean deviate from the mode in skewed distributions Mean is most sensitive to skew Therefore mean should be used ONLY in normal distributions as a measure of central tendency
42
Quantification of Variation of the Distribution Pattern: Range
Highest to lowest value of the distribution – 58 to 75 in illustration No measure of how much each value deviates from point of central tendency
43
Quantification of Variability: Variance
Assess deviation of each score from the mean – deviation score Sum of these deviation scores = 0 so take sum of deviation scores squared – called the sum of squares (SS) Divide the SS by the number of scores (n) to get the SAMPLE VARIANCE also called mean square (MS) Better estimation of the POPULATION VARIANCE is to divide the sum of squares by n-1
44
Measuring Variability & Distribution Pattern
Standard Deviation (s or SD) is derived from the mean value Mean is only meaningful in normal distribution so mean & SD should be used only with normal distributions Mode is best measure of central tendency & range is a good measure of variability in skewed distributions
45
Quantification of Variability: Standard Deviation (s or SD)
Equals the square root of the variance (s2) Has a meaning in normal distributions –± 1 SD represents 68% of total # values in distribution –± 2 SD represents 95% of total # values –± 3 SD represents 99.7% of total # values Excellent discussion of variance & standard deviation can be found at: http://www.mathsisfun.com/data/standard-deviation.html
46
Percentile
Value represents what percent of the total number of values in a data set In a data set of 40 values – the 10th value from the lowest in the set is the 25th percentile The median value (20th of 40 values) is the 50th percentile Commonly reported quartiles, quintiles or deciles * Quartiles - each group comprises a quarter of the entire data set * Quintiles – ranked data set are the four points that divide the data set into five equal groups * Deciles – ranked data set are the nine points that divide the data set into ten equal groups
47
Quartiles
Quartiles of a ranked data set are the three points that divide the data set into four equal groups - 25th percentile, 50th percentile & 75th percentile Each group comprises a quarter of the entire data set Different Quartiles: * Lower (bottom) quartile – \< 25th% * Lower middle quartile − 25th% − 50th% * Upper middle quartile − 50th% − 75th% * Upper (top) quartile − \> 75th% Interquartile range is from 25th% − 75th%
48
Graphic Display of Mean/SD
Common display in graphs of mean as a point and an error bar showing one SD around that mean Allows for easy comparison between groups
49
Graphic Display of Median, Quartiles & Range (estimate)
Box & Whisker Plot: * Line at center of box - Median * Box - Interquartile range (IQR) * Whiskers (bars) - 1.5 x IQR – approximates range Box & Whisker plots allow for easy comparison between groups when using median, quartiles and ranges
50
Coefficient of Variation (CV) –CV = (SD/mean) (100)
Independent of units (cancel out) so can compare distributions using different units Expresses SD as a proportion to the mean which account for differences in the mean so is a relative variation
51
Standardized Score (z-score)
Z-score = Mean divided by SD ((X)/SD) Normalizes mean to the variability of the data set
52
What do inferential statistics allow us to do?
Allow investigators to make decisions from sample data about the larger population
53
What is probability? Give an example with coins & dice.
Likelihood that an event will occur given all possible outcomes Single Coin Toss: * 2 sides with both having an equal chance of coming face up when tossed * So probability = 0.5 Probability of Single Die Toss * Six sides so probability = 0.167
54
What does probability mean in terms of an individual or population?
Applied to Groups of Data The greater the deviation from mean the greater the probability of NOT being in the group Critical when comparing two groups – what is the probability of 2 groups being from different populations * 68% of all values are within 1 SD of mean * 95% of all values are within 2 SD of mean * 99% of all values are within 3 SD of mean * \> deviation from mean less probability of being in the group
55
Probability of Pair of Dice Toss
value of 2 - 1/36 – p=.028 * value of 3 – 2/36 – p=.056 * value of 4 – 3/36 - p=.083 * value of 5 – 4/36 - p=.111 * value of 6 – 5/36 - p=.139 * value of 7 – 6/36 - p=.167 * value of 8 – 5/36 - p=.139 * value of 9 – 4/36 - p=.111 * value of 10 – 3/36 - p=.083 * value of 11 – 2/37 – p=.056 * value of 12 – 1/36 - p=.028
56
What is sampling error?
In statistics, sampling error is the error caused by observing a sample instead of the whole population. The sampling error is the difference between a sample statistic used to estimate a population parameter and the actual but unknown value of the parameter
57
# Define the standard error of the mean.
The standard error of the mean (SEM) (i.e., of using the sample mean as a method of estimating the population mean) is the standard deviation of those sample means over all possible samples (of a given size) drawn from the population. The smaller the standard error, the more representative the sample will be of the overall population.
58
What is the effect of sample size on standard error of the mean?
The standard error is inversely proportional to the sample size; the larger the sample size, the smaller the standard error because the statistic will approach the actual value.
59
What is experimental error?
Experimental error is the difference between a measurement and the true value,or between two measured values. Within group variance due to experimental error or error variance Experimental error unexplained variation including personal differences between individuals, methodological inconsistencies, behavior differences between individuals & environmental factors Not mistakes just unexplained variation
60
Distributions of Outcome Measure
Applied to Groups of Data Distributions of outcome measure before (blue) & after (red) an intervention What probability that the intervention produced a change in the outcome? What probability that the difference between these two distributions is due to chance alone and not the result of the intervention?
61
Hypothesis Testing
Difference between group before and after intervention produces two hypotheses: Null Hypothesis: * No difference between the pre & post- intervention groups * No treatment effect * Any difference just due to chance * Any difference is functionally 0 * NOT SIGNIFICANT difference Alternative – Experimental Hypothesis * Difference between pre & post groups is large enough so that the difference is not due to chance alone * Functional difference * SIGNIFICANT difference
62
Experimental Errors
Reject or not reject the experimental or null hypothesis – 4 possible outcomes ## Footnote Correctly reject null hypothesis – experimental hypothesis is true Correctly reject experimental hypothesis – null hypothesis is true Incorrectly reject null hypothesis – null hypothesis is true, called Type I error Incorrectly reject experimental hypothesis – experimental hypothesis is true – Type II Error
63
Type I Error
Incorrect rejection of null hypothesis – null hypothesis is true Example: * Decide that a treatment works when it really doesn’t * Early studies of radical mastectomy versus lumpectomy reported superiority of radical mastectomy in certain forms of breast cancer * Incorrectly as it turned out – so many women had unnecessary disfiguring surgery
64
Type II Error
Incorrectly reject experimental hypothesis – experimental hypothesis is true – Type II Error Example: * Decide that a treatment does not work when it really does * Investigators wish to avoid this to enhance chance for publication
65
Type I Error – Levels of Significance
Level of significance – alpha (α) level – is set before beginning a study This is the risk the investigator is willing to take in making a Type I error Usual risk\* is considered being wrong 1 in 20 times, a 5% error rate or an α = 0.05 Can make it harder or easier to make a type I error If set risk at 1 in 100 times, a 1% error rate or an α = 0.01 – harder to make Type I error If set risk at 1 in 10 times, a 10% error rate or an α = 0.1 – easier to make Type I error When it is made harder to make a Type I error it becomes easier to make a Type II error – so there is a trade off
66
Type II Error – Levels of Significance
Type II Error – Statistical power Incorrect rejection of the experimental hypothesis Probability of making a type II error is beta (β) β= 0-1-0.2 are commonly accepted levels Which is the probability or likelihood that we will be unable to determine statistically significant differences 1-β is the statistical power of a test –is the probability that the test can determine a significant difference if it exists \> power means \> ability to determine a significant difference 0.8-0.9 is 80-90% 80% are commonly used levels
67
Levels of Significance
Significance is determined by using a statistical test and it provides a “p” (probability) level When p \< 0.05, the test shows a significant difference between group – that this difference did not occur by chance alone When p ≥ 0.05, the test shows that there was not a significant difference between groups – that this difference could have occurred by chance alone Once significance level is set cannot change for that study Dichotomous decision – significant or not significant, never say “highly significant” or “more significant” * If statistical test shows a p \<.001 you can have a high confidence that there is a significant difference, it is not “highly significant” * If p=0.04 you may have less confidence in the significant difference
68
4 Determinants of Statistical Power
Significance criterion Variance of the data Sample size Effect size
69
Significance Criterion
α & β are logical, not mathematical, opposites Risk of type I error when the risk of type II error so peril in changes
70
Variance of Data Set
Increase power of statistical test when data variance decrease When variability within groups increase harder to tell difference so decrease power Variance decrease & power increase by: * Using repeated measures * Homogeneous subjects * Controlling sources of random measurement error Sometimes methodological changes can be done to reduce the variance of the data set – change population used or exclusion criteria Increasing size of the sample – more potential values clustered around mean and less influence of outliers
71
Sample size
Small samples are * Less likely to reflect the population therefore decreases true differences between groups * Increases effect of outliers so increases variability * Decreases effect size THIS IS THE PRINCIPLE WAY TO INCREASE STATISTICAL POWER IN A STUDY TO AVOID A TYPE II ERROR
72
Effect size
Factor that assesses true difference between groups If means of 2 groups very different but groups have a high variability – there may not really be a difference between the 2 groups If there is small difference between the means of two groups, but each group is very homogeneous, then could have a true difference between the groups Effect size involves variability within groups & mean difference between groups Effect size equals difference between means of two groups divided by the pooled standard deviation Cohen states that ES \< 0.2 is small, around 0.5 is medium & \> 0.8 is large Power Analysis is done before the study (a priori) to determine the needed sample sizes
73
Non-Directional versus Directional Hypothesis
Are the two groups just different or are they different in a particular way – one group \> or \< other Non-directional hypothesis – states only that there is a difference and does not predict which is \> or \< Directional hypothesis – predicts which is \> or \< Statistical Tests must reflect directionality or non-directionality of hypothesis
74
1-tailed vs 2-tailed tests
2-tailed test used for non-directional hypotheses * 2-tailed test the critical region of α = 0.05 is divided between upper & lower part of the curve so in each region α/2 1-tailed test used for directional hypothesis * 1-tailed test the entire critical region is in the region in the direction of the predicted difference so full critical region of α = 0.05 in that direction 1-tailed test more powerful & less likely to make a Type II error, but more likely to make a type I error 1-tailed test reflects placement of full critical region in one direction
75
Degrees of Freedom
Degree of freedom = # values -1 5 numbers which add up to 30 – if you change the values of 4 of those to add up to 30 the fifth can only have a preordained value and NOT one you can change So 5 numbers have 4 degrees of freedom * Used in most statistical calculations
76
Parametric Statistics
Used to make inferences about a population Use requires assumption of a Normal distribution Large populations of interval or ratio (metric) data of often normally distributed “Goodness of fit” or “homogeneity of variance” tests used to determine normality of distribution Non-parametric tests when assumption of normality not met Parametric tests have \> statistical power so prefer to use those if possible
77
Statistical versus Clinical Significance
Statistics are tools and should not be substituted for clinical knowledge or judgment There may be a statistically significant change in say ROM which has no clinical implication – that is not producing a functional difference Conversely, a sizable change in a number of subjects can be produced, but because of variability of the result, there is not a significant difference Not enough # of subjects to avoid a type II error Responders versus non-responders – non-responders in the experimental group will act like control group So with subgroups may have a true difference Important considerations in interpreting statistical results: * What do your numbers mean? – understand the phenomenon * Don’t confuse statistical significance with size of treatment effect – statistical significance is only the probability of change NOT a reflection of the size of the effect What is a meaningful change? * Don’t confuse significance with a meaningful change - a statistically significant change may not be a meaningful change
78
What comprises between group & within group comparisons?
Within group differences is the experimental error Between group difference includes both intervention + experimental error In RCTs the random assignment should equally distribute the characteristics that give rise to experimental error between the two groups; therefore should not contribute to between group comparisons Relative magnitude of between group differences & within group variability give rise to overlap or not between groups Two groups not different –reject experimental (alternative) hypothesis Two groups different – reject null hypothesis
79
What gives rise to overlap or not between groups?
Relative magnitude of between group differences & within group variability give rise to overlap or not between groups
80
What is the basis for the statistical analysis of the two-group difference?
Ratio of between groups to within groups–Between groups difference has both treatment effect + experimental error Within groups variability due sole to experimental error * So if no treatment effect the ratio is 1 Desired outcome is a \>\> a number as possible when treatment effect \>\>\>\> experimental error Think of it like a signal to noise ratio – lot of signal compared to noise wanted
81
What is the meaning of the ratio of between groups to within group differences?
Within group differences is the experimental error Between group difference includes both intervention + experimental error Compare between group differences to within group variability to calculate a T value and compare it to the critical T-value (T-value for p \< 0.05) to determine significance
82
What differences and uses of T-test for independent groups (unpaired) & for dependent groups (paired)
Developed by William Gossett – Guinness Brewery –pseudonym was Student, so often called Student T-test Compare between group differences to within group variability to calculate a T value and compare it to the critical T-value (T-value for p \< 0.05) to determine significance Comes in two forms: * Unpaired T-test - for independent groups * Paired T-test - for dependent groups One & Two-tailed Tests * 1-tailed versions of both T-tests used with directional comparisons * 2-tailed versions of both T-tests used with non-directional comparisons * So need to use both terminologies when describing what test is used
83
What is the problem of using multiple T-tests?
(Used for two group comparisons, so if more than two groups being compared would have to use multiple T-tests) Use of multiple T-tests (or any test for that matter) higher risk of committing a Type I error For \> 2 group comparisons use the general analysis of variance (ANOVA) T-test is a specialized form of ANOVA for 2 group comparisons only
84
How do you use confidence intervals to test for null hypothesis?
Look at the interval between the two means and potential variations of that interval if the study was repeated If the possible variation of intervals between the two means includes 0 that indicates there is the potential for no difference – and the experimental hypothesis can be rejected
85
What is the rationale for using T-test versus CI?
T-test shows significant difference CI demonstrates statistical significance but also estimates the size of the effect as well
86
Unpaired T-test
Independent because two different groups of individuals with two different levels of within group variability Use to compare control group with experimental group
87
Paired T-test
Dependent because it compares two measurements for the same group of people with very similar within group variability because they are the same people Test compensates for the similarity of variances Use includes comparison of pre-post changes in a single group