Gradient Descent Flashcards

1
Q

What is gradient descent?

A

It is an optimization algorithm used to minimize a function, the function we are talking about is the lost function ( we defined it in the previous lesson).
It is fundamental for training ML models .
The formula is a bit of a mess so we view it on iPad

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is the main hypothesis we use when we compute gradient descent,

A

The function has to be differentiable , (in stochastic process we are not able to calculate the derivative)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Why is it usefull the taylor series in computing the gradient descent?

A

Can be used to approximate the derivative which are computanionally intensive to calculate

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is the main definition of gradient descent

A

So the gradient is the generalization of the derivative in more variables

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is the consequence of considering **alpha ** too large or too small in the gradient descent

A

If alpha is too large we have a time of convergence faster but it isn’t guarantee while in the other case the convergence is surely going to happen but in a slow time

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Write down the formula and explain it

A

So the formula is messy to so I would do this exercise on paper + explaining

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Given a function f(x,y), x0, alpha ; write down the iteration to do (exam question)

A

Resolution exercise on paper

How well did you know this?
1
Not at all
2
3
4
5
Perfectly