ResNets Flashcards

(24 cards)

1
Q

What problem do ResNets aim to solve?

A

Vanishing and exploding gradients in deep neural networks.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What causes vanishing gradients in deep networks?

A

Repeated multiplication of small derivatives during backpropagation.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is a common sign of gradient instability in deep models?

A

Abnormal gradient distributions, such as near-zero or spiked values.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What are three signs of unstable gradient flow?

A

Abnormal gradients, chaotic learning curves, and irregular layer outputs.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Why is ReLU preferred over sigmoid in deep networks?

A

ReLU better preserves gradients during backpropagation.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What is the core idea behind residual connections?

A

Instead of learning y = f(x), learn y = f(x) + x.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is a residual block?

A

A network unit that adds its input to its output after a series of transformations.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Why do residual connections help with training deep models?

A

They allow gradients to flow more easily through the network.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What does f(x) + x mean in a ResNet?

A

The output is the sum of the learned transformation and the original input.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What happens if f(x) learns nothing in a ResNet?

A

The identity connection ensures the network can still pass input forward.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What analogy links ResNets and LSTMs?

A

Both preserve information over structure—LSTM across time, ResNet across depth.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

How do residual connections relate to vanishing gradients?

A

They reduce the chance of vanishing gradients by providing an unimpeded gradient path.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What are skip connections in ResNets not equivalent to?

A

Encoder-decoder skip paths like in U-Nets.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What does a typical ResNet block contain?

A

Two convolutional layers and a skip connection with optional batch norm and ReLU.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What is one advantage of using residual blocks in CNNs?

A

They allow very deep networks to be trained effectively.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What is ResNet-34?

A

A 34-layer residual network designed for ImageNet-level performance.

17
Q

What architecture enabled the training of 100+ layer CNNs?

A

Residual Neural Networks (ResNets).

18
Q

What is one limitation of ResNets?

A

Reduced interpretability due to multiple forward paths.

19
Q

What can happen if residual blocks are poorly designed?

A

They may default to identity mappings and learn nothing useful.

20
Q

What is a computational cost of using ResNets?

A

Increased parameter count and training time due to added layers.

21
Q

Why might debugging ResNets be difficult?

A

Because of the complexity introduced by residual pathways.

22
Q

What do residual blocks encourage the network to learn?

A

Only the difference (residual) between input and desired output.

23
Q

What paper introduced ResNets?

A

‘Deep Residual Learning for Image Recognition’ by He et al., 2015.

24
Q

Why is deeper not always better in plain CNNs?

A

Deeper networks can suffer from degraded training due to gradient issues.