Week 9 - Neural Networks Flashcards

1
Q

What are the two types of learning a neural network can do?

A

Supervised and unsupervised

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is supervised learning?

A

The output of the neural network is compared against the correct output. The network then corrects itself based on that output. Overall, the data is labelled

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is unsupervised learning?

A

The data is not labelled. The network organises itself according to patterns in the data and no external “desired output” is provided

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What does a perceptron consist of?

A

It consists of a set of weighted connections, the neuron (incorporating the activation function) and the output axon

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

How does a perceptron learn?

A

Initialise weights & threshold

Present the input and desired output

Calculate the actual output of the network:
For each input:
* Multiply the input data (xi) by its weight (wi).
* Sum the weighted inputs and pass through the activation function
Adapt the weights:
* If correct wi(t+1) = wi(t)
* If output 0, should be 1 wi(t+1) = wi(t)+xi(t)
* If output 1, should be 0 wi(t+1) = wi(t)-xi(t)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What can the weight update function be used to give?

A

A decimal term η between 0.0 and 1.0 to slow learning

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What does the Widrow-Hoff Learning Rule give us?

A

Weight updates proportionate to the error made, giving Δ = desired output – actual output

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Name a limitation of the perceptron.

A

Only linearly separable problems can be solved

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What does linearly seperable mean?

A

A straight line which seprates two classes can be drawn

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What are the three layers of an MLP (Multi-Layer Perceptron)?

A

Input, hidden, output

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What are weights in terms of neural networks?

A

Variable strength connections between units that propagate signals from one unit to the next. They are the main component changed during learning

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Describe the feedforward learning algorithm.

A

Initialise weights and thresholds to small random values.

Present input and desired output.

Calculate actual output by:
- Multiplying incoming signal by weight
- Pass this through sigmoid activation function
- Pass on this output to units in the next layer

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Describe the backpropagation learning algorithm.

A

Adapt the weights.

Start from the output layer and work backwards:
- New weight (t+1) = old weight, plus a learning rateerror for pattern p on node joutput signal for p on j

Compute error as follows:
- For output units, compute error sigmoid derivativetarget output - actual output
- For hidden units, use the sigmoid derivative
weighted error of the k units in the layer above

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What are the two types of weight updating?

A

Batch and online

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Describe batch weight updating.

A

All patterns are presented, errors are calculated, then the weights are updated

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Describe online weight updating.

A

The weights are updated after the presentation of each pattern

17
Q

What does momentum do in terms of weight updating?

A

Encourages the network to make large changes to weights if the weight changes are currently large. This allows the network to avoid local minima in early stages as it can overcome hills

18
Q

List some differences between symbolic AI and connectionism?

A

Symbolic:
explicit representation (symbols)
expert system (IF-THEN) rules
enabling reasoning
knowledge programmed (by humans)
serial (fragile)
does not generalise (outside scope)
understandable/explainable

Connectionism:
implicit representation (numbers)
neural network (weighted graph)
enabling perception
knowledge learned (from data)
distributed (graceful degradation)
generalise (outside scope)
black-box

19
Q

Name some properties of neural networks.

A

Able to learn to relate input variables to required output

Able to generalise between samples

Shows graceful degradation

20
Q

Define classification.

A

The output of the function to learn is a class (discrete)

21
Q

Define regression.

A

The output of the function to learn is a value (continuous)

22
Q

What is graceful degradation?

A

In symbolic systems, the removal of one component of the system usually results in failure

Removal of neuron(s) from a neural network will reduce performance and probably not result in overall failure, which replicates out understanding of fault tolerance in the brain (graceful degradation)

23
Q

What is classification designed to do?

A

Group samples according to some known properly

24
Q

What are the two datasets required for a classification problem?

A

Training (consists of a set of measurements (inputs) and a class (output), used for learning) and testing (“unseen” examples to test generalisation)

25
Q

What is an issue in representing continuous data in a neural network?

A

Can require normalisation depending on the activation function

26
Q

What is an issue in representing discrete data in a neural network?

A

Each value needs to have a separate representation in the network to avoid implied order bias (noise that can confuse a neural network)

27
Q

What are the two representations of discrete data in a neural network?

A

Field-type and thermometer-type

28
Q

When does overfitting occur and how can it be overcome?

A

When the network is trained on a task for too long, the network learns the noise in the input as well as common patterns, so the result is poor on unseen examples

It can be overcome by stopping the network learning earlier - early stopping

29
Q

What is a common method for early stopping?

A

Cross-validation - have three data sets: training, testing and cross-validation