week 4 - chatgpt Flashcards

(11 cards)

1
Q

What limitation of single-layer perceptrons is addressed by multilayer neural networks?

A

Single-layer perceptrons cannot model nonlinear decision boundaries; multilayer networks can model arbitrary nonlinear functions.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is the universal approximation theorem in neural networks?

A

A feedforward network with a single hidden layer can approximate any continuous function given sufficient neurons and proper weights.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Why are differentiable activation functions needed in multilayer neural networks?

A

Because backpropagation relies on gradient descent, which requires the derivative of the activation function to compute updates.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

How does backpropagation compute the weight updates in a neural network?

A

By using the chain rule to propagate the error backward from the output layer through the hidden layers and updating weights using gradient descent.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is the general update rule for weights using backpropagation?

A

w ← w − η * ∂J/∂w, where J is the cost function and η is the learning rate.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What is the error term for an output unit in backpropagation?

A

δ_k = (t_k − z_k) * f′(net_k), where t_k is the target and z_k is the output.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is the error term for a hidden unit in backpropagation?

A

δ_j = f′(net_j) * sum over k of (w_kj * δ_k), where δ_k is the error of output neurons.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is the main purpose of using stochastic (online) backpropagation?

A

To update weights after each training example, which often leads to faster convergence and better generalisation than batch updates.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What problem does early stopping help prevent during training?

A

It helps prevent overfitting by stopping training when validation error starts to increase, even if training error is still decreasing.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

How is a Radial Basis Function (RBF) network different from a Multilayer Perceptron (MLP)?

A

An RBF network uses radial activation functions based on distance and typically has only one hidden layer, while an MLP uses layered linear combinations with nonlinear activations.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What are the two main training phases of an RBF network?

A

First, determine the centres of the basis functions (unsupervised), then learn output weights (supervised).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly