Chapter 15 Flashcards

(23 cards)

1
Q

What type of data are RNNs designed to process?

A

Sequential or time-series data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is a recurrent neuron?

A

A neuron that receives feedback from its previous output to influence future outputs.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

How is an RNN trained?

A

Using backpropagation through time (BPTT), unrolling the network across time steps.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is the “unrolling” of an RNN?

A

Representing the same RNN layer at multiple time steps to visualize flow across time.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What are the two main weight matrices in a recurrent neuron?

A

One for the current input (Wx) and one for the previous output (Wy).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What is a memory cell in RNNs?

A

A structure that preserves state over time, helping the network retain information.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is the sequence-to-sequence architecture?

A

A model that takes a sequence as input and produces a sequence as output (e.g., time-series forecasting).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is a sequence-to-vector model?

A

A model that takes a sequence input and produces a single output (e.g., sentiment analysis).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What is a vector-to-sequence model?

A

A model that takes a single input and generates a sequence (e.g., image captioning).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What is an encoder-decoder model in NLP?

A

A model that encodes an input sequence to a vector and decodes it into an output sequence (e.g., translation).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is the difference between naive forecasting and deep learning for time series?

A

Naive forecasting predicts the last value again; deep models learn patterns for better accuracy.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What is MSE and why is it used?

A

Mean Squared Error; a loss function used to measure prediction accuracy in time series.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What is the drawback of predicting one step at a time in time series forecasting?

A

Error accumulation over successive steps.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What is the advantage of predicting multiple steps at once in time series forecasting?

A

Reduced error accumulation and more stable training gradients.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What is the main issue when handling long sequences in RNNs?

A

Unstable gradients and memory loss of earlier inputs.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What is batch normalization and how does it help RNNs?

A

Normalization across batches; it can stabilize training but is difficult to apply across time.

17
Q

What is layer normalization in RNNs?

A

Normalization across feature dimensions, easier to use than batch normalization for RNNs.

18
Q

What is the short-term memory problem in RNNs?

A

RNNs forget inputs after many time steps due to vanishing gradients.

19
Q

What is an LSTM cell?

A

A memory cell with gates (forget, input, output) that maintains long-term memory in sequences.

20
Q

What is a GRU cell?

A

A simplified LSTM with fewer gates, merging hidden and cell states.

21
Q

How do 1-D convolutional layers work in sequence modeling?

A

They apply filters across time steps to detect temporal patterns.

22
Q

What is WaveNet and what does it do?

A

A deep neural network for sequence generation using dilated convolutions to capture long-range dependencies.

23
Q

What is the role of dilation in WaveNet?

A

It allows the network to look back further in time without increasing the number of layers.