Lecture 12 - Deep Learning Flashcards

(34 cards)

1
Q

What are the three layers in CNNs?

A

Input layer
Hidden layer
Output layer

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Do we need all the edges to be connected?

A

No, info can be shared, as you can represent small regions with few parameters, ie beak example

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is a CNN?

A

A neural network with some convolutional layers.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is a Convolutional Layer?

A
  • A convolution layer has a number of filters that does convolutional operations
  • The filters are known as the parameters to be learned
  • These are 3x3 filters with specific patterns
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

How does the convolutional layer (filter) work?

A

REFER TO SLIDES

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What is the understanding of convolution using a line?

A

It basically uses a line, it detects the line with the maximum value

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

How does convolution extract and learn features?

A

REFER TO SLIDES

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

When is padding necessary in convolution?

A

Padding is used if you want the same output ie the original matrix is 6x6 but as you can see we get a 4x4 image, so we need to add padding if we want a 6x6 and then do the same procedure

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

How does convolution work for RGB images?

A

It’s the same thing for colour, except you will have 3 layers, 1 for each colour and each filter needs to be done 3 times (one for each colour)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What is the difference between Convolution and Fully Connected Layer?

A
  • Convolution: We don’t take all the values we only focus on a small area
  • Fully connected layer: all the features connected to the next neuron, so in the example the first neuron will get all 36 features (shown by X36)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What are shared weights?

A

REFER TO SLIDES

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What is the entire CNN process?

A
  • This is the process of CNN you start with the input image, you perform convolution and max pooling as many times are required, you perform flattening, this then gives you the full CNN
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What is Max Pooling?

A

Max pooling is basically the feature map is split into even sections using a filter, and then the max value in each section is taken
- In other words we are keeping the most prominent features that are found through convolution using max pooling

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What is average pooling?

A

There is also average pooling, which is taking all the values in a given sections adding it up and dividing it by the total number of points

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Why do we use pooling (max pooling)?

A
  • Max pooling is a form of subsampling and down sampling, as we saw before the image went from 4x4 (the feature map) to 2x2 (the sections created from the feature map)
    ○ This means we have fewer parameters
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What are the ways CNN compresses a fully connect network?

A

Reduce the number of connections
Having shared weights on the edges

17
Q

What are the steps for max pooling?

A

These are the full steps leading to smaller image
- We take the big image
- Perform convolution
- Perform max pooling
- Get smaller image (where each filter is a channel - the filter being the layers as you can see this is 2 x 2 x 2)

18
Q

What is Relu?

A
  • An activation function that occurs after each convolution
  • What happens is we only keep the positive values
  • Given by m(0, x) which basically means if value is not positive it becomes 0, if it is positive keep it
19
Q

What is flattening?

A

Take the features and convert it into a column, which is then converted into the full connected feedforward network

20
Q

What is the pipeline for Deep Learning-based Computer vision?

A

Input -> Deep Learning for feature extractionm description, classfication/regression/segmentation -> Output

21
Q

What do CNN Filters Learn?

A
  • They learn full objects i.e. faces, car, etc (high-level features)
  • Parts such as nose, eyes, car parts, legs of chair (mid-level features)
  • Edges (low-level features)

As you go down the layers become more fine grain

22
Q

What is Normalisation?

A

Take the value multiply it by the mean and divide it by the standard deviatio

23
Q

Why do we do batch normalisation?

A
  • If inputs are not centered around 0
  • If inputs have different scaling per element
24
Q

What does batch normalisation do?

A
  • Makes the inputs have zero mean and unit variance
  • To this you apply the equation
    ○ Take the value multiply it by the mean and divide it by the standard deviation
25
What are the different types of normalisation?
- Batch: computed across the batch (vary in batch size) - Layer: across the layer - Instance: Applied on a single instance - Group: Gets a group of instances and apply normalisation
26
What are the Example CNNs?
LeNet5 AlexNet VGGNet GoogleNet ResNet
27
What is the LeNet5 Example?
- Had an architecture of [conv, pool, conv, pool, fully connect layer, fully connected layer] - Providing 4 layers - Where: ○ Conv filters were 5x5 at stride 1 ○ Pool was 2x2 at stride 2
28
What is the AlexNet Example?
An example of shallow network CNN - Architecture is [conv, pool, norm, conv, pool, norm, conv, conv, conv, pool, Fully connected, Fully connected, Fully connected] - Providing 8 total layers - Where: ○ Conv filters were 11 x 11 at stride 4 ○ Pool filters were 3x3 at stride 2
29
What is the VGGNet Example?
Example of deep network CNN - Different between VGG16 and VGG19 is the number of layers VGG19 has 3 more - Uses a conv of 3x3 stride 1 ○ Why?: has the same effect as a 7x7 as well as allows it to go deeper, more non-linearities and fewer parameters
30
What is GoogleNet Example?
Example of deeper network CNN with computation efficiency - Have 22 layers - Conv use 1x1 filters - Pool uses 3x3, but it uses average pooling not sampling (max pooling)
31
What is ResNet Example?
An example of revolution of depth - Has 152 layers - Why was this possible? ○ Using vanishing gradient, where the gradient basically becomes very small after some time (gradient finishes) ... They use skip connection or identity connection, they take the previous convulsion and add it to the current one or in other words take the previous gradient and add it to the current gradient
32
What should the deeper model learn to be at least as good as the shallower model?
Copying the learned layers from the shallow model and setting adddtional layers for mapping
33
What layers learn in CNNs?
Convolution Final CNN model - Provide features maps or final model
34
What layers dont learn in CNNs?
Max Pooling Relu Flattening - They just perform transformation