deep neural network Flashcards

(28 cards)

1
Q

What defines a deep neural network (DNN)?

A

A neural network with more than one hidden layer.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is the main difference between shallow and deep networks?

A

Deep networks stack multiple nonlinear layers to build complex representations.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is the key operation performed by each layer in a DNN?

A

A linear transformation followed by a nonlinear activation.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is the formula for the pre-activation in a layer?

A

z = W·h + b

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What does the activation function do in each layer?

A

Applies nonlinearity to the pre-activation output.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What is the final output of a deep network?

A

The result after all layers have transformed the input.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

How can you mathematically express a DNN’s output?

A

ŷ = f_L(…f_2(f_1(x)))

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Why do we stack layers in a DNN?

A

To learn hierarchical and compositional features.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What do early layers typically learn in deep networks?

A

Low-level features (e.g., edges or color blobs).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What do deeper layers typically learn?

A

High-level patterns (e.g., object parts or abstract concepts).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is the geometric intuition of depth in DNNs?

A

Each layer folds space, allowing complex reshaping of input regions.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What is the benefit of composing multiple shallow functions?

A

It enables the model to build complex mappings from simple components.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What happens to input space with each ReLU in a deep network?

A

It gets partitioned into more piecewise linear regions.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What is the effect of depth on model expressiveness?

A

It allows exponential growth in the number of linear regions.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What is an activation region in a DNN?

A

A subspace where a specific set of ReLU units are active.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

How does parameter count compare between deep and shallow networks for similar region complexity?

A

Deep networks may use fewer parameters to achieve the same complexity.

17
Q

What is one advantage of deep networks over shallow ones?

A

They can learn more abstract and reusable features with fewer units.

18
Q

What is the typical architecture of a DNN layer?

A

Linear → Activation → Output → Next layer input

19
Q

What hyperparameter controls the number of layers in a DNN?

A

The depth (number of hidden layers).

20
Q

What hyperparameter controls the size of each layer?

A

The number of hidden units (width).

21
Q

What activation function is most commonly used in DNNs?

A

ReLU (Rectified Linear Unit).

22
Q

What is the universal approximation theorem?

A

A shallow network can approximate any continuous function with enough units.

23
Q

Why do we still prefer deep networks despite the universal approximation theorem?

A

Deep networks can approximate functions more efficiently and with better generalisation.

24
Q

What does each layer in a DNN build upon?

A

The representation produced by the previous layer.

25
What is the main benefit of hierarchical feature learning?
It allows reuse of features across different parts of the input.
26
What is the notation for hidden layer output in DNNs?
h^{(l)} = φ(W^{(l)} h^{(l-1)} + b^{(l)})
27
What is a key challenge in deep networks?
Training and optimisation due to non-convex loss surfaces.
28
What is one way DNNs are more efficient than shallow networks?
They reuse parameters through depth to represent complex functions.