CNN Architectures and Applications Flashcards

1
Q

What is the primary problem with Fully Connected Networks (FCNs) regarding image processing?

A

Images are flattened into 1D vectors, losing spatial structure

This loss of spatial structure makes it difficult for FCNs to effectively understand visual data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

How do Convolutional Neural Networks (CNNs) address the issues of FCNs?

A

Preserve spatial locality and handle images more intelligently

CNNs maintain the spatial relationships in images, improving accuracy.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is a Convolutional Neural Network (CNN)?

A

A deep learning model that works especially well with visual data like images

CNNs are specifically designed to process pixel data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is the key idea behind how CNNs function?

A

They learn patterns in an image using filters

Filters help in detecting features such as edges and textures.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is the convolution operation in CNNs?

A

A small matrix called a kernel slides over the image, multiplying and summing to create a feature map

This operation extracts features from the input image.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What does the term ‘stride’ refer to in the context of CNNs?

A

How far the filter moves each step

Stride affects the size of the output feature map.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What are the two types of padding in CNNs?

A

‘valid’ (no padding) and ‘same’ (padding with zeros)

Padding is used to control the spatial dimensions of the output.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is the function of the convolutional layer in a CNN architecture?

A

Extracts feature maps

Convolutional layers are crucial for identifying patterns in images.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What does the pooling layer do in a CNN?

A

Shrinks size while keeping important information

Pooling helps reduce the computational load and overfitting.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What is the purpose of the activation layer in a CNN?

A

Adds non-linearity (e.g., ReLU)

Activation functions enable CNNs to learn complex patterns.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is the role of the fully connected layer in CNNs?

A

Final decision-making (e.g., classification)

Fully connected layers integrate features learned by previous layers.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

How do CNNs learn?

A

By adjusting filters to minimize prediction error

Learning occurs through a process called backpropagation.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What is the basic process of CNN learning?

A

Input image goes through convolution + pooling, output passed to fully connected layers, loss function calculates error, backpropagation computes gradients, optimizer updates weights

This process iteratively improves the model’s predictions.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What is an example of the input and output of a convolutional layer?

A

Input: 32×32 image, Output: 30×30×16

The output dimensions depend on the kernel size and stride.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What are the total learnable parameters in a CNN example provided?

A

255,632

The number of parameters affects the model’s capacity to learn.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What is max pooling in CNNs?

A

Picks the biggest number in a region

Max pooling helps retain the most significant features.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

What is average pooling in CNNs?

A

Takes the average of numbers in a region

Average pooling provides a smoother representation of features.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

What are the benefits of pooling layers in CNNs?

A
  • Reduces size
  • Controls overfitting
  • Speeds up training

Pooling layers are essential for efficient training and model performance.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

What happens in the fully connected layer of a CNN?

A

All features combine to make a decision

This layer synthesizes information from all previous layers.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

What do convolutional layers learn using?

A

Filters

Filters are crucial for feature extraction.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

What is the purpose of pooling in CNNs?

A

Downsample while keeping key info

Pooling reduces dimensionality, making the model more efficient.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

What does activation introduce in CNNs?

A

Non-linear twist (usually ReLU)

Non-linear activations allow CNNs to model complex functions.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

What is the role of fully connected layers in CNNs?

A

Combine all for final output

Fully connected layers determine the final classification or regression output.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
24
Q

What is the function of filters (kernels) in convolutional layers?

A

Scan over input image and create feature maps

Filters are essential for identifying patterns in the data.

25
What is stride in the context of convolutional layers?
How far the filter moves ## Footnote Stride impacts the output size of the feature map.
26
What does 'valid' padding mean?
No padding, smaller output ## Footnote Valid padding reduces the spatial dimensions of the feature map.
27
What does 'same' padding do?
Keeps input/output sizes the same (zero-padding edges) ## Footnote Same padding allows for consistent output sizes.
28
What is the typical structure of each convolutional layer?
Convolution and Activation (e.g., ReLU) ## Footnote This structure is fundamental for feature extraction.
29
What is the most common pooling method used in CNNs?
MaxPooling2D ## Footnote Max pooling is widely used due to its effectiveness in feature retention.
30
What is the pool size and stride for MaxPooling2D?
Pool size: (2,2), stride: (2,2) ## Footnote This configuration is standard for reducing dimensions.
31
What do fully connected layers do with learned features?
Flatten them and feed into final classifier ## Footnote Flattening transforms the multi-dimensional output into a single vector.
32
What is LeNet known for?
An early CNN architecture for digit recognition (MNIST) ## Footnote LeNet laid the groundwork for many modern CNN architectures.
33
What is the key idea behind LeNet architecture?
Alternating Conv → ReLU → Pool ## Footnote This sequence is fundamental for effective feature extraction.
34
What do feature maps in LeNet use to save memory?
Shared weights ## Footnote Sharing weights reduces the number of parameters in the model.
35
What is the training configuration for the LeNet model?
20 epochs, Batch size: 128, Optimizer: Adam ## Footnote This configuration is typical for training CNNs.
36
What are two regularization techniques mentioned for improving generalization?
* Spatial Dropout * Hyperparameter Tuning ## Footnote These techniques help prevent overfitting and improve model performance.
37
What problem can arise during CNN training?
Vanishing Gradients or Exploding Gradients ## Footnote These issues can hinder the training process significantly.
38
What does batch normalization do?
Makes training smoother, faster, and more stable ## Footnote Batch normalization helps in stabilizing the learning process.
39
What is the formula for batch normalization?
Norm = (x - mean) / sqrt(var + ε) Out = γ * Norm + β ## Footnote This formula standardizes the input for each mini-batch.
40
What are the benefits of using batch normalization?
* Reduces sensitivity to weight init * Slightly regularises model (less overfitting) * Allows higher learning rates ## Footnote Batch normalization enhances the overall performance of the model.
41
When is it recommended to add batch normalization in Keras?
Before the activation ## Footnote Placing batch normalization before activation functions can lead to better results.
42
What do convolutional layers do in a CNN?
Extract features ## Footnote Convolutional layers are vital for identifying key patterns in the input data.
43
What is the purpose of pooling layers?
Shrink spatial size ## Footnote Pooling layers help reduce computational complexity.
44
What is the function of fully connected layers?
Make decisions ## Footnote Fully connected layers synthesize information from previous layers to output final predictions.
45
What do activation functions add to a neural network?
Non-linearity ## Footnote Non-linear activation functions are crucial for learning complex relationships.
46
What does batch normalization do during training?
Stabilises training ## Footnote Batch normalization helps in maintaining consistent learning dynamics.
47
What is transfer learning?
Reusing a pretrained model for your own task ## Footnote Transfer learning leverages existing knowledge to improve learning efficiency.
48
What are the benefits of transfer learning?
* Saves time * Works well with small datasets * Leverages existing learned features ## Footnote Transfer learning is especially useful in scenarios with limited data.
49
What is one strategy for transfer learning?
Freeze layers and only retrain the final classifier ## Footnote This strategy helps retain learned features while adapting to new tasks.
50
What is data augmentation?
Pretend you have more data by slightly changing existing images ## Footnote Data augmentation helps improve model robustness.
51
List examples of data augmentation techniques.
* Rotate * Flip * Zoom * Brightness tweak * Crop ## Footnote These techniques enhance the diversity of the training dataset.
52
What does dropout do during training?
Randomly 'turn off' neurons ## Footnote Dropout helps prevent overfitting by ensuring that the model does not rely too heavily on any individual neuron.
53
What problem does dropout help prevent?
* Overfitting * Reliance on specific paths in the network ## Footnote Dropout encourages a more generalized model.
54
What is a common value for dropout?
Dropout(0.5) → 50% of neurons are dropped ## Footnote This value is often used to balance model complexity and performance.
55
What is a characteristic of VGG architecture?
Uses only 3×3 convolutions ## Footnote VGG architecture is known for its simplicity and depth.
56
What notable achievement is associated with AlexNet?
Won ImageNet 2012 ## Footnote AlexNet marked a significant advancement in deep learning for image classification.
57
What is a key feature of ResNet architecture?
Uses skip connections (identity mappings) ## Footnote Skip connections help in training very deep networks by mitigating the vanishing gradient problem.
58
What does ReLU stand for and what does it do?
ReLU: max(0, x); Fast and simple activation function ## Footnote ReLU is widely used due to its efficiency and effectiveness in learning.