Recurrent Neural Networks Flashcards

(12 cards)

1
Q

What kind of data are RNNs designed to process?

A

Sequential data such as speech, music, text (sentences), DNA sequences, and time series.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Why do we need recurrence (feedback loops) in neural networks for sequences?

A

To remember past inputs and share learned features across time steps, enabling handling of variable-length inputs and context.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is a vanilla RNN’s key characteristic in its computational graph?

A

It has directed cycles (feedback loops) that propagate hidden states over time.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Name four typical RNN use-case structures on input/output length.

A

Many-to-one (e.g., sentiment classification), one-to-many (image captioning), many-to-many (machine translation), and synchronized many-to-many (video classification).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is the vanishing gradient problem in RNNs?

A

Gradients diminish exponentially over long sequences, making it hard for vanilla RNNs to learn long-term dependencies.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What are the two gates in a Gated Recurrent Unit (GRU)?

A

The update gate and the reset gate, which control memory content and state updates.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

How does a GRU update its hidden state?

A

It uses the update gate to combine the previous state with a candidate state computed using the reset gate.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What are the three gates in an LSTM cell?

A

Input gate, forget gate, and output gate, which regulate cell state writing, resetting, and reading.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Why are LSTMs more complex to train than GRUs or vanilla RNNs?

A

They have more gates and parameters, increasing computational cost and requiring more data to learn effectively.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Compare vanilla RNN, GRU, and LSTM in terms of training difficulty and effectiveness.

A

Training difficulty: RNN < GRU < LSTM; effectiveness: RNN < GRU ≈ LSTM.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Give two real-world examples where LSTMs have been successfully applied.

A

Google Translate for speech translation; Facebook’s daily automatic translations; Apple’s QuickType and Siri text prediction.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What major architecture supplanted RNNs for sequence tasks and why?

A

Transformers replaced RNNs because they enable parallel sequence processing and avoid vanishing gradients via self-attention.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly