Linear Models Flashcards
Recall the standard linear regression model, define any terms and indices
yi=B0+B1xi1+B2xi2+…+Bkxik+ei
yi is the ith response
xij is the ith value of the jth regressor
k is the number of regressors (k+1=p parameters)
B0 is the y-intercept
Bj is coefficient associated with the jth regressor
ei is an error term, usually assumed iid random with mean 0 and variance sig^2
Matrix form of regression model and what are any assumptions made (OLS assumptions)?
y = XB + e
E[e]=0
var(e)=sig2 I
OLS estimate of B variance and expected value
Bhat = (X’X)-1X’Y
Var(Bhat)= sig2(X’X)-1
E(Bhat)=B
What does OLS estimate, Bhat, do?
Minimizes the sum of squared residuals (in matrix form: SSres= [y-yhat]’ [y-yhat])
SSres in quadratic form (Y’AY where A is symmetric)
Y’[I - X(X’X)-1X’]Y
Global F-test
H0: B1=B2=…=Bk=0
Ha: Bj ≠ 0
Test statistic is MSreg / MSres ~ F
Where MSreg = SSreg / k
SSreg = SUM (yhati - ybar)2
Write SSreg in terms of matrix notation
SSreg = Y’ [X(X’X)-1X’ - 1(1’1)-11’]Y
Describe a cell means model
yij= ui + eij
yij is the jth observation for the ith group
i=1,2,…,t
j=1,2,…,ni
ui is the true mean for the ith group
eij is the error term
For a cell means model, define y, M, and u
What is the incidence matrix and what is uhat, M’M, (M’M)-1?
Less than Full Rank ANOVA Model (overly paramterized)
Where the row length of X is larger than the column length
Difference between overly parameterized and fully parameterized
Overly paramterized models have columns in the X matrix which are linear combinations of the first column (ie not orthogonal)
For less than full rank anova model, what happens to Bhat?
Since Bhat= (X’X)-1X’Y it cannot be estimated because X’X does not have an inverse. But, X’X does have generalized inverses which have to be solved that way
Consequences of Less than Full rank ANOVA models
Generalized inverses are not unique
There are no unique estimates for B
Serious limitations on what we can estimate and test
Usual focus: contrasts (pairwise comparisons, factorial effects)
Things of critical importance: eigenvalues and rank of matrices (provides degrees of freedom)
If A and B are nxn square matrices then det[AB]=
det[A]det[B}
If A is nxn then det[A]=0 iff
A is singular
The rank of A is…
the greatest number of linearly independent columns (or rows) of A. (Linear dependence implies that at least one column (or row) of A can be written as a linear combination of the other columns (or rows) )
If A and B are non-singular, then for any matix C
C, AC, CB, ACB
all have the same rank
If A is an mxn matrix of rank r, then there exist non-singular matrices P and Q such that PAQ is one of the following:
The rank of AB cannot
exceed the rank of either A or B
If A is a nxn matrix then det[A]=0 iff
the rank of A is less than n
The matrix of a quadratic form can always
be chosen to be symmetric
A and B are said to be congruent matrices iff
there exists a non-singular matrix, C, such that
B=C’AC
We say C is the congruent transformation of A
Let A be an nxn symmetric matrix of rank r. There exists a non-singular matrix C such that
C’AC=D
where D is a diagonal matrix with exactly r non-sero diagonal elements