3rd year Flashcards

(29 cards)

1
Q

In a generalised linear model, how is the mean of the response variable related to the linear
predictor?

A

In a GLM, the mean response µ is related to the linear predictor η according to
g(µ) = η,
where g(·) is the link function.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

How is the variance related to the linear predictor?

A
The variance of the response is related to the linear predictor η through the mean µ as
Var{y} = V (µ)a(φ) = V (g^−1(η))a(φ),
where V (µ) is the variance function.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Explain what the canonical link is

A

The canonical link is the same function of µ as θ is (a function of µ).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

The Newton-Raphson algorithm becomes identical to which

algorithm named after a famous statistician when the canonical link is used?

A

The Newton-Raphson algorithm becomes identical to the Fisher scoring algorithm when the canonical
link is used.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Explain the role of the link function g(µ) in a generalised linear model

A

The role of the link function g is to transform the mean response µ so that g(µ) = η, the linear
predictor.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Write down the general form of a distribution from the exponential family

A

f(y) = exp

[(yθ − b(θ))/a(φ)] + c(y, φ)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Explain how a generalised linear model is more general than a linear model with normal errors

A

A generalised linear model is more general in two ways.
(i) The distribution of the response variable can be any member of the exponential family.
(ii) The mean response µ need not be a linear function of the explanatory variables, as long as g(µ)
is, for some link function g

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Write down the log linear regression model with Poisson responses yi ∼ Pois(λi) on a single explanatory
variable xi with intercept, i = 1, . . . , n, stating any further assumptions on y1, . . . , yn

A

Poisson response log linear model with intercept:
yi = µi + εi ∼ Pois(µi) independent,
log(µi) = β0 + β1xi
, i = 1, . . . , n.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q
Explain what the updating equation
β(k+1) = (X′W(k)X)^(−1)X
′W^(k)ξ^(k)
, k = 0, 1, 2, . . .
does
A

The updating equation calculates β^(k+1) as weighted least squares estimates of regression coefficients, using working responses ξ^(k) with predictor values in X and weights in W(k) on the diagonal.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Explain the connection between Poisson regression and logistic regression with binomial responses.

A

In large scale studies of rare events a Poisson regression with log link and offset gives nearly the same
results as logistic regression

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Write down the linear regression model as a GLM for yi ∼ N(µi, σ2) on a single explanatory variable x with values xi, i = 1, 2, . . . , n and state any further assumption on y1, y2, . . . , yn. An
intercept or constant term should be included in the model.

A

As a GLM, the linear regression model with normal errors is given by
yi = µi + εi ∼ N(µi, σ2) independent, µi = β0 + β1xi, i = 1, . . . , n.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Give three reasons why a Poisson model can be used to analyse contingency table data.

A
(i) It has been shown in class that
yij ∼ Pois(λij )|Pij yij = n ∼ Multinomial with πij = λij/Pij λij.
(ii) The MLE problems are equivalent.
(iii) An additive Poisson model with log link is equivalent to independence between row and column
classifications.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

The F-distribution with n1 and n2 degrees of freedom is defined as

A

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

In the full (saturated) model

A

all the θi
’s are free to vary, so that the fitted value
for yi equals yi
, i = 1,…,n

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

The scaled deviance of a GLM

A

is twice the difference in maximum loglikelihood

value when comparing it with the full (saturated) model.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

The variance function is

A

V (µ) = b′′(θ) written as a function of µ

17
Q

What is the likelihood equation in matrix form?

A

The likelihood equation in matrix form is
X′W (ξ − Xβ) = 0
where X = (xi j) is the n × p design/data matrix, W is a diagonal matrix with
wi =1/[V(µi) . g′(µi)^2], i = 1,…,n
on the diagonal line, and ξ is a column vector with elements
ξi = x′iβ + g′(µi)(yi −µi), i = 1,…,n.

18
Q

What are the projection errors y − yˆ

called

19
Q

How is the deviance related to the residual sum of squares? How do you estimate
σ^2 using the deviance?

A

The deviance equals the residual sum of squares: D = SSE.

σ^2 is estimated by σˆ2 = SSE/(n − p) = D/(n − p).

20
Q

For f (y;θ,ϕ) in the exponential family, the Fisher information in y about θ is

A

i(θ) =b′′(θ)/a(ϕ)

21
Q

Properties of ξi

A
E[ξi] = x′_iβ 
Var{ξi} = g′(µi)^2 . V(µi)a(ϕ)  = w^(−1)_i . a(ϕ)
22
Q

The scaled deviance is

23
Q

Coefficient of determination

A

R^2 = SSR/SST

24
Q

The logit link is preferred to the probit link Φ−1(π) because

A
  • it provides a canonical link within the framework of a GLM;
  • it makes it easy to compute the parameter estimates,
  • it has interpretation in terms of odds ratio.
25
w_i =
=1/[V(µi) . g′(µi)^2
26
ξ_i =
= x′_iβ + g′(µi)(yi −µi)
27
The scaled deviance of a GLM is
twice the difference in maximum loglikelihood | value when comparing it with the full (saturated) model.
28
logit link g (µ) =
log µ/(1−µ)
29
The iteratively re-weighted least squares algorithm is given by the updating equations
β^k = (X′.W^k.X)^(−1) . X′.W^k.ξ^k