recurrent neural networks Flashcards

(25 cards)

1
Q

What is the key idea behind Recurrent Neural Networks (RNNs)?

A

They maintain a hidden state to remember information across time steps.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What type of data are RNNs designed for?

A

Sequential or time-series data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Why is a feedforward network unsuitable for time-dependent inputs?

A

It treats all inputs as independent and ignores temporal order.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

How does an RNN incorporate memory?

A

By passing a hidden state from one timestep to the next.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What does the formula aₜ = f(Wxₜ + Uaₜ₋₁ + b) represent?

A

The update rule for the RNN hidden state.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What is parameter sharing in RNNs?

A

Using the same weights at every timestep to process inputs consistently.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is the benefit of parameter sharing?

A

Reduces the number of parameters and improves generalisation.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What activation function is commonly used in basic RNNs?

A

Tanh or ReLU.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What does ‘many-to-one’ RNN architecture mean?

A

A sequence of inputs produces a single output.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What does ‘many-to-many’ RNN architecture mean?

A

A sequence of inputs produces a sequence of outputs.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is the main limitation of standard RNNs during training?

A

They suffer from vanishing and exploding gradient problems.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What causes the vanishing gradient problem in RNNs?

A

Repeated multiplication by weights less than 1 during backpropagation.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What causes the exploding gradient problem in RNNs?

A

Repeated multiplication by weights greater than 1, causing large gradients.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What is the impact of vanishing gradients on learning?

A

Prevents learning of long-term dependencies.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What is the impact of exploding gradients?

A

Leads to unstable updates and divergence during training.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What does backpropagation through time (BPTT) do?

A

Unrolls the network across time and computes gradients through each step.

17
Q

In an RNN, what role does the weight wₐ play in memory?

A

It controls how much past information is carried forward.

18
Q

What happens if the memory weight wₐ is 0.5 over 2 steps?

A

The signal quickly diminishes (e.g., 10.5 with input 7).

19
Q

What happens if the memory weight wₐ is 2.0 over 2 steps?

A

The signal explodes (e.g., 168 with input 7).

20
Q

What kind of real-world task was used to motivate RNNs in this lecture?

A

Rainfall prediction using radar image sequences.

21
Q

What is the key advantage of RNNs over feedforward networks for sequences?

A

They model temporal dependencies through recurrent connections.

22
Q

What is the initial hidden state a₋₁ usually set to?

23
Q

Why is unrolling the RNN necessary for training?

A

To apply gradient-based optimisation across all timesteps.

24
Q

How does an RNN process inputs over time?

A

Sequentially, one timestep at a time, updating the hidden state.

25
What kind of function is used to compute the output in an RNN?
A non-linear function of the current input and previous hidden state.