chatGPT questions exam2 Flashcards

1
Q

What does Stochastic Gradient Descent (SGD) optimize in machine learning models?

A

SGD is used to minimize the loss function of a model.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

How does SGD differ from traditional gradient descent?

A

Unlike traditional gradient descent that uses the entire dataset for updates, SGD updates model parameters using only a single sample or a small batch of samples.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What are the key benefits and drawbacks of using SGD?

A

SGD is more computationally efficient for large datasets and can help escape local minima, but it may lead to slower and less stable convergence.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What does the FROC curve evaluate in medical image analysis?

A

the FROC curve evaluates detection and localization performance, plotting sensitivity against the average number of false positives per image.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Why can accuracy be a misleading performance metric?

A

Accuracy can be misleading in cases of class imbalance, where it doesn’t reflect the model’s performance on the minority class

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What was the major limitation of the original R-CNN in object detection?

A

The original R-CNN was slow due to its reliance on selective search for generating region proposals.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

How did Fast R-CNN improve over R-CNN?

A

Fast R-CNN improved efficiency by sharing computations across region proposals.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What innovation did Faster R-CNN introduce to object detection?

A

Faster R-CNN introduced a Region Proposal Network (RPN), allowing for end-to-end training and faster proposal generation.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

How does setting a high threshold for predictions affect sensitivity and specificity?

A

A high threshold generally increases specificity but may decrease sensitivity by rejecting true positives with lower confidence.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What approach do convolutional networks use for segmentation tasks?

A

They utilize architectures with downsampling and upsampling layers, and sometimes strided or dilated convolutions, to produce a pixel-wise segmentation map.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

How is the number of trainable parameters in a convolutional layer calculated?

A

The formula is (filter height × filter width × input channels + 1) × number of filters, where “+1” accounts for the bias term.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Why is the learning rate considered a critical hyperparameter in neural network training?

A

t influences the training dynamics, where too high a rate may cause divergence and too low a rate results in slow convergence.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What does dropout do in neural networks?

A

Dropout prevents overfitting by randomly dropping units and their connections during training, simulating training multiple networks in parallel.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What are the advantages of ReLU over sigmoid functions?

A

ReLU mitigates the vanishing gradient problem, accelerates convergence, and maintains gradient flow for positive inputs.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Why is feature normalization important before training a machine learning model?

A

it scales features to have zero mean and unit variance, ensuring consistent scaling and preventing information leakage from the test set.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

How do residual networks (ResNets) facilitate the training of deep models?

A

resNets use skip connections to learn residual mappings, alleviating the vanishing gradient problem and enabling deeper architectures.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

What does the receptive field refer to in convolutional networks?

A

It refers to the size of the input area that influences the network’s output, determined by the cumulative effect of convolution and pooling operations.

18
Q

How does the softmax function work in neural networks?

A

It converts raw output scores into probabilities by taking the exponential of each output and normalizing these values by the sum of all exponentials.

19
Q

What does a false negative (FN) indicate in detection tasks?

A

A FN occurs when an object is present but not detected by the classifier, which is critical in applications where missing detections can have serious consequences.

20
Q

What is a dilated (atrous) convolution and its purpose?

A

It increases the receptive field without increasing the number of weights by inserting spaces between kernel elements, allowing for broader spatial aggregation.

21
Q

How does U-Net architecture specialize for medical image segmentation?

A

U-Net uses a symmetric architecture with downsampling and upsampling paths, enabling precise localization and context integration without needing a pre-defined weight map.

22
Q

How is specificity calculated in a model’s performance evaluation?

A

Specificity is calculated as the number of true negatives divided by the sum of true negatives and false positives.

23
Q

What determines the output size of a valid convolution operation?

A

The output size is determined by (W − F + 1) × (H − F + 1), where W and H are the width and height of the image, and F is the filter size.

24
Q

What makes the YOLO object detection framework unique?

A

YOLO processes the entire image in a single evaluation and makes predictions for each grid cell, combining bounding box predictions and class probabilities.

25
How does backpropagation work in training neural networks?
It computes gradients of the loss with respect to network parameters by applying the chain rule, allowing for efficient parameter updates.
26
What are histogram-based features in image processing?
These features, like mean, variance, skewness, and kurtosis, describe the distribution of pixel intensities and are invariant to/do not consider spatial relationship and correlation between pixels (identical histograms can belong to different textures)
27
how does the shift-and-stitch algorithm benefit fully convolutional networks?
It enables the generation of fine-grained outputs by reconstructing the original resolution through multiple shifted input passes.
28
What improvement does Fast R-CNN offer over the original R-CNN?
Fast R-CNN processes the entire image at once and uses RoI pooling to extract features, improving efficiency and speed
29
How does histogram matching adjust an image's brightness and contrast?
It modifies an image so that its histogram matches that of a reference image, using cumulative distribution functions for mapping pixel values.
30
What are the implications of not using pooling layers in CNNs?
Omitting pooling layers requires convolutional layers with strides for downsampling, potentially leading to higher memory and computational costs.
31
How does backpropagation facilitate learning in CNNs?
It calculates gradients (using chain rule) for all network parameters by propagating errors backward through the network, enabling parameter updates for learning.
32
What are the steps in mini-batch gradient descent training?
The steps are selecting a mini-batch, performing a forward pass, computing loss, backpropagating to compute gradients, and updating weights.
33
How does the choice of pooling size and stride affect max-pooling operations?
Incompatible pooling sizes and strides may lead to suboptimal coverage and downsampling uniformity in the feature map.
34
What strategy is used in transfer learning with small datasets?
Freeze most of the pre-trained network's parameters and fine-tune the last few layers to adapt the features to the new task.
35
How does dilated convolution affect the receptive field in CNNs?
dilated convolution expands the receptive field without increasing the number of weights, allowing for broader spatial information aggregation.
36
What is the purpose of the softmax function in neural networks?
it converts logits into probabilities, facilitating multi-class classification by ensuring output values sum to 1.
37
What does the term "dying ReLU" problem refer to?
it refers to the issue where neurons stop learning due to always outputting zero, which can stem from improper initialization or gradient descent issues. - this can be solved with leakyrelu
38
How is the number of trainable parameters in a dense layer calculated?
It is calculated as (number of inputs + 1) × number of outputs, accounting for each input connection and bias term.
39
What does zero padding do in convolutional layers?
Zero padding allows control over output dimensions, maintaining spatial sizes or adjusting them for specific convolutional effects.
40
How is the "best model" identified during neural network training?
The best model is often the one with the lowest validation loss, indicating effective learning and generalization capabilities without overfitting.