week 5 - chatgpt Flashcards

1
Q

What is the main advantage of using deep neural networks over shallow ones?

A

Deep networks can represent complex functions with fewer parameters by exploiting hierarchical features, making them more efficient for complex tasks.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is the vanishing gradient problem in deep networks?

A

As gradients are backpropagated through many layers, they can become very small, preventing early layers from learning effectively.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is the exploding gradient problem in deep networks?

A

Gradients can grow exponentially as they are backpropagated, leading to unstable updates and divergence during training.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

How does the ReLU activation function help with vanishing gradients?

A

ReLU has a constant derivative for positive inputs, preventing the gradient from shrinking to zero during backpropagation.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is the difference between ReLU, LReLU, and PReLU?

A

ReLU outputs 0 for negative inputs; LReLU uses a small fixed slope; PReLU learns the slope from data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Why is weight initialization important in deep networks?

A

Proper initialization (e.g., Xavier or He) maintains variance across layers, avoiding vanishing or exploding activations and gradients.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is the purpose of batch normalization?

A

It standardizes layer inputs to have zero mean and unit variance, which stabilizes and speeds up training.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

How do skip connections help train deep networks?

A

They allow gradients to bypass some layers, mitigating the vanishing gradient problem and enabling very deep architectures.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What is the main idea behind convolutional neural networks (CNNs)?

A

CNNs use local filters and weight sharing to detect spatial hierarchies in data, such as edges, textures, and shapes in images.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What role do pooling layers play in CNNs?

A

Pooling layers reduce spatial dimensions and introduce translation invariance, helping to generalize better across input variations.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is the purpose of 1x1 convolutions in CNNs?

A

They are used to reduce or expand the number of channels without affecting spatial dimensions, aiding computational efficiency.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What is dropout and why is it used?

A

Dropout randomly deactivates neurons during training, forcing redundancy and improving generalization by reducing overfitting.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly