Neural Networks and Deep Learning Foundations Flashcards

1
Q

What is the main difference between Traditional Machine Learning and Deep Learning?

A

Traditional Machine Learning involves manually selecting features, while Deep Learning learns features automatically from raw data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is the purpose of Gradient Descent in machine learning?

A

To reduce errors in predictions and improve the model’s accuracy.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What are the steps involved in the Backpropagation process?

A
  • Forward Pass
  • Compute Loss
  • Backpropagation
  • Gradient Descent
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

True or False: Neural Networks are inspired by the structure of the human brain.

A

True

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What are the three types of layers in a Neural Network?

A
  • Input Layer
  • Hidden Layers
  • Output Layer
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Fill in the blank: The equation for a prediction in a neural network is Prediction = Input × Weight → _______.

A

[Activation Function]

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What are the two types of Loss Functions mentioned?

A
  • Binary Cross-Entropy
  • Categorical Cross-Entropy
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is the role of an Optimizer in deep learning?

A

To help the AI learn faster by adjusting learning rates and strategies.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What dataset is used in the example of Handwritten Digit Recognition?

A

MNIST Dataset

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What is the first step in building a simple neural network using Keras?

A

Import Libraries

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is the purpose of normalizing pixel values in the MNIST dataset?

A

To improve training efficiency by scaling values from 0-255 to 0-1.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What is One-Hot Encoding used for in the context of neural networks?

A

To convert labels into a format the neural network can understand.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What activation function is used in the first layer of the example neural network?

A

ReLU (Rectified Linear Unit)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What metric is used to evaluate the performance of the model on test data?

A

Accuracy

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

True or False: Keras simplifies the process of building neural networks.

A

True

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What is the goal of training a neural network on the MNIST dataset?

A

To correctly predict handwritten digits (0-9).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

What is the goal of Backpropagation and Gradient Descent?

A

Minimize the error between predicted and actual outputs.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

What do activation functions prevent in neural networks?

A

They prevent neural networks from behaving like linear regression models and allow them to learn complex relationships.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

What is the equation without activation functions?

A

output = dot(W, input) + b

20
Q

What is the equation with activation functions?

A

output = ReLU(dot(W, input) + b)

21
Q

What is a Linear Activation Function?

A

Output = Input (Straight line)

22
Q

What is the main issue with the Sigmoid activation function?

A

Vanishing Gradient – When values go beyond ±3, the gradient becomes tiny, and learning slows down.

23
Q

What does the Softmax activation function do?

A

Converts values into probabilities that sum to 1.

24
Q

What is an example of Softmax output?

A
  • Cat: 70%
  • Dog: 20%
  • Bird: 10%
25
What are the benefits of the Tanh activation function?
Maps inputs between -1 and 1, allowing negative values.
26
What is a key problem with Tanh?
Still suffers from vanishing gradient for large values.
27
What distinguishes ReLU from other activation functions?
Only activates for positive values; negative inputs become 0.
28
What is a problem associated with ReLU?
Dead Neurons – if a neuron only gets negative values, it stops learning.
29
How does Leaky ReLU address the dead neuron problem?
Gives negative values a small value instead of zeroing them out.
30
What is the best use case for the Sigmoid activation function?
Probability outputs (binary classification).
31
What problems are associated with Softmax?
Can be computationally expensive.
32
When is Tanh best utilized?
In situations where negative values are needed.
33
What are the common uses for ReLU?
Deep learning models (default choice).
34
What is the main advantage of Leaky ReLU?
Prevents dead neurons and improves training.
35
Fill in the blank: Activation functions allow _______ models to learn complex data.
[deep learning]
36
What is the most commonly used activation function for hidden layers?
ReLU
37
What activation function is best for multi-class classification?
Softmax
38
What is the purpose of activation functions?
They enable deep networks to learn complex data relationships.
39
What is a key limitation of linear activation functions?
They can't handle complex patterns.
40
Name a common activation function used for probabilities.
Sigmoid
41
What is the main drawback of the sigmoid activation function?
It suffers from vanishing gradients.
42
Which activation function is used in multi-class classification?
Softmax
43
What range does the Tanh activation function map inputs to?
-1 and 1.
44
What is the most commonly used activation function?
ReLU
45
What problem does Leaky ReLU address?
The 'dying neuron' problem.