Flashcards in Chapter 2 Deck (25):
the probability of making a Type I error (usually this value is .05).
the probability of making a Type II error (Cohen, 1992, suggests a maximum value of 0.2).
Central limit theorem
this theorem states that when samples are large (above about 30) the sampling distribution will take the shape of a normal distribution regardless of the shape of the population from which the sample was drawn. For small samples the t-distribution better approximates the shape of the sampling distribution. We also know from this theorem that the standard deviation of the sampling distribution
for a given statistic calculated for a sample of observations (e.g. the mean), the confidence interval is a range of values around that statistic that are believed to contain, with a certain probability (e.g. 95%), the true value of that statistic (i.e. the population value)
Degrees of freedom
an impossible thing to define in a few pages let alone a few lines. Essentially it is the number of 'entities' that are free to vary when estimating some kind of statistical parameter. In a more practical sense, it has a bearing on significance tests for many commonly used test statistics (such as the F-ratio, t-test, chi-square statistic) and determines the exact form of the probability distribution for
the difference between the observed value of a variable and the value of that variable predicted by a statistical model.
an objective and (usually) standardized measure of the magnitude of an observed effect. Measures include Cohen's d, Glass' g and Pearson's correlations coefficient, r.
how sexually attractive you find a statistical test. Alternatively, it's the degree to which a statistical model is an accurate representation of some observed data. (Incidentally, it's just plain wrong to find statistical tests sexually attractive.)
a model that is based upon a straight line.
this is a statistical procedure for assimilating research findings. It is based on the simple idea that we can take effect sizes from individual studies that research the same question, quantify the observed effect in a standard way (using effect sizes) and then combine these effects to get a more accurate idea of the true effect in the population.
a test of a directional hypothesis. For example, the hypothesis 'the longer I write this glossary, the more I want to place my editor's genitals in a starved crocodile's mouth' requires a one-tailed test because I've stated the direction of the relationship (see also two-tailed test).
in statistical terms this usually refers to the collection of units (be they people, plankton, plants, cities, suicidal authors, etc.) to which we want to generalize a set of findings or a statistical model.
the ability of a test to detect an effect of a particular size (a value of .8 is a good level to aim for).
a smaller (but hopefully representative) collection of units from a population used to determine truths about that population (e.g. how a given population behaves in certain conditions).
the probability distribution of a statistic. We can think of this as follows: if we take a sample from a population and calculate some statistic (e.g. the mean), the value of this statistic will depend somewhat on the sample we took. As such the statistic will vary slightly from sample to sample. If, hypothetically, we took lots and lots of samples from the population and calculated the statistic of
the extent to which a statistic (e.g. the mean, median, t, F, etc.) varies in samples taken from the same population.
an estimate of the average variability (spread) of a set of data measured in the same units of measurement as the original data. It is the square root of the variance.
the standard deviation of the sampling distribution of a statistic. For a given statistic (e.g. the mean) it tells us how much variability there is in this statistic across samples from the same population. Large values, therefore, indicate that a statistic from a given sample may not be an accurate reflection of the population from which the sample came.
Standard error of the mean (SE)
the full name of the standard error.
Sum of squared errors (SS)
another name for the
a statistic for which we know how frequently different values occur. The observed value of such a statistic is typically used to test hypotheses.
a test of a non-directional hypothesis. For example, the hypothesis 'writing this glossary has some effect on what I want to do with my editor's genitals' requires a two-tailed test because it doesn't suggest the direction of the relationship. (See also One-tailed test.)
Type I error
occurs when we believe that there is a genuine effect in our population, when in fact there isn't.
Type II error
occurs when we believe that there is no effect in the population when, in reality, there is.