Week 4: Multilayer Perceptrons and Backpropagation Flashcards

1
Q

Feedforward Neural Networks

A

There’s 1 input layer, hidden layer(s), and 1 output layer. Each layer is an array of neurons, with layers interconnected by links. Each link has a connection weight. The connections are one-way, going forward.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Activation/Transfer Function

A

Activation functions take on various forms, with the main purpose of transforming the input into a desirable range of outputs.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Symmetric Hard Limit (Signum) Transfer Function

A

f(n) = sgn(n)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Linear Transfer Function

A

f(n) = n

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Symmetric Sigmoid Transfer Function

A

f(n) = 2/(1+e^{-2n}) - 1

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Logarithmic Sigmoid Transfer Function

A

f(n) = 1/(1+e^{-n})

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Radial Basis Transfer Function

A

f(n) = e^{-n^2}

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Backpropagation

A

This method allows weights to be updated based on the effective error for each hidden unit using gradient descent.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Learning Curves

A

These curves measure the training, test, and validation error over the training process.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Radial Basis Function Neural Networks

A

The input unit uses the linear transfer function. The hidden unit use the Radial Basis Function. The output unit can use any activation function.

d_j = ||x - c_j||, with x being inputs, c_j being the centres of the basis function. c_j can be preset or determined by a training algorithm. The result of the hidden node is h(d_j).

Similarities with MLP Networks:
- Both are universal approximators
- Both are nonlinear feed-forward neural networks

Differences with MLP Networks:
- RBF networks has only 1 hidden layer, but MLP networks can have more than 1 hidden layer.
- RBF networks use different basis functions from activation functions used in MLP networks.
- RBF networks compute distance between input patterns and centres while MLP networks compute inner product of input patterns and weights.
- RBF are trained with a 2-phase algorithms but MLP networks are trained with a single-phase algorithm

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Basis Functions

A

These functions are used in the hidden nodes. As d_j approaches infinity, h(d_j, \sigma) approaches 0.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Gaussian Function

A

h(d_j) = e^{\frac{-d_j^2}{ 2 \sigma ^2}}, \sigma > 0

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Multi-quadric Function

A

h(d_j) = (d_j^2 + \sigma^2)^{0.5}, \sigma > 0

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Generalised Multi-quadric Function

A

h(d_j) = (d_j^2 + \sigma^2)^{\beta}, \sigma>0, 0 < \beta < 1

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Inverse Multi-quadric Function

A

h(d_j) = (d_j^2 + \sigma^2)^{-0.5}

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Generalised Inverse Multi-quadric Function

A

h(d_j) = (d_j^2 + \sigma^2)^{-\beta}, \sigma>0, \beta > 0

17
Q

Thin Plate Spline Function

A

h(d_j) = d_j^2 \ln(d_j)

18
Q

Cubic Function

A

h(d_j) = d_j^3

19
Q

Linear Function

A

h(d_j) = d_j