Statistical methods Flashcards
(39 cards)
What is the chain rule of probability?
P( X = x n Y = y ) = P( Y = y ) x P( X = x | Y = y )
or
P( X = x n Y = y ) = P( X = x ) x P( Y = y | X = x )
What is P( X = x | Y = y ) using the chain rule of probability
P( X = x n Y = y ) / P( Y = y)
If X and Y are independent, what do we know?
P( X = x | Y = y ) = P( X = x )
If X and Y are independent, what is P( X = x n Y = y ) ?
P( X = x ) x P( Y = y )
What is covariance?
Measures how two random variables vary together
What is the law of total probability
P( X = x ) = Sum of P( X = x n Y = y)
What is variance?
Measures the average deviation of a variable away from its expected value
If a new random variable: Z = X + Y, then what is var(Z)
var (Z) = var (X) + var (Y) + 2cov (X,Y)
What is correlation?
tells you how close the data is to lying on a straight line
If we are in a binomial distribution, and we are trying to approximate P( X >= n), what do we need to do?
Convert to the normal distribution and use the continuity correction
If we are in a normal distribution and want P( X > n ), what do we need to do?
Standardise using Z = (X - mean) / standard deviation
What is the sample mean?
1/n x (sum of Xi)
An estimator, mu hat, is unbiased if:
E ( mu hat) = mu
What is the size of the bias
E (mu hat) - mu
If we assume observations are iid, what assumptions have we made?
E (Xi) = Mu x
var (Xi) = sigma^2 x
cov (Xi,Xj) = 0 for all i,j
What is the central limit theorem
Provided n is sufficiently large, the sample mean (x bar) would be approximately distributed centred around the expectation E(X), with variance var(X) / n
When do we use the z-distribution
When n is large and we know the population variance
What is the sample variance
(1 / n - 1) x sum of (Xi - X bar)^2
When is an estimator considered ‘good’
If it is unbiased
If it is efficient (low variance)
What is the size of the error when estimating the population parameter
Error = Mu hat - Mu
What is the average error
E (Mu hat - Mu) = E (Mu hat) - Mu
What is the mean square error
MSE (Mu hat) = var (Mu hat) + bias^2 (Mu hat)
When is an estimator consistent in terms of MSE
lim (n to infinity) MSE (Mu hat) = Mu
When is an estimator consistent (Weak law of large numbers)
plim (Mu hat) = E (X) = Mu x