Linear Models for Regression Flashcards

1
Q

What is regression?

A

The problem in which we have a set of points from a function and we want to aproximate that function without knowing the original function.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is the relationship between regression and classification?

A

Any regression problem can be blassed as a classification one. Each point for regression lies on the decision boundary (we just need to string them together)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What are the 3 basis for finding a function?

A

> Polynomial basis

> “Gaussian” basis

> Sigmoid basis

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is polynomial basis?

A

ϕi(x) = xi

y(x,w) = w0 + w1x + w2x2 + … + wMxM

Set ϕ0 = 1

y(x,w) = w0 + ∑ (wiϕi(x)) = wTϕ(x)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is Gaussian basis?

A

ϕi = e^(-(x - μi )2 / (2s2 ))

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What is the sigmoid basis?

A

ϕi(x) = σ((x - μ) / s) = σ(u,s)

σ(a) = 1 / (1 + e-a)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is the vector equation that we would use to calculate the exact solution to a regression problem and what is required for this?

A

w = Φ-1 t

This requires that matrix, Φ, be square so we can only pick a square number of points.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What does overdetermined mean? and what impact does this have?

A

This is when there are more equations than varaibles and there is no exact solution possible. Instead we can define an error to minise this.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What is the equation for error?

A

E = 0.5 ∑(ϕiTw - ti )2

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Derrive the equations for the least squares solution

A

E = 0.5 ∑ (ϕiTw - ti )2

∇E = ∑ ϕiiTw - ti)

This will be minimum when the gradient ∇E = 0

ϕT(ϕw - t) = 0

ϕTϕw = ϕTt

w = (ϕTϕw)-1 ϕTt

ϕp = (ϕTϕw)-1 ϕT

w = ϕpt

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is the data set and equation form and what are the steps for sum of squares solution?

A

Data set:

  • (x, t)
  • (x_value, target)
  • Equation:
  • Example: y = w0 + w1f(x)

Basis:

  • Polynomial
  • Gaussian
  • Sigmoid

Step 1: Calculate ϕ using the basis and bias

ϕ = [bias, basis]

[… , …]

Step 2: Compute ϕTϕ

Step 3: Compute (ϕTϕ)-1

Step 4: Compute ϕp = (ϕTϕ)-1ϕT

Step 5: Compute the result using the targets w = ϕpt

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

How can sequential learning be applied to the sum of squares solution? Why do we want to do this? What is this process called?

A

With lots of points, this is a computationally expensive process so we can consider one point at a time and use gradient descent. This is the least-mean-squares algorithm

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What is the equation for the least mean squares algorithm?

A

wt+1 = wt - ηϕnTnw - tn)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Why is the sum of squares error equation ideal?

A

Because it is convex and has one global minimum

How well did you know this?
1
Not at all
2
3
4
5
Perfectly