Test 5 Flashcards
(8 cards)
5 Step Training Loop
- Generate Prediction - model(x, y)
- Compute Loss - loss(y, y~)
- Zero Gradient Parameters (optimizer.zero_grad())
- Compute Gradients (loss.backward())
- Change weights - optimizer.step()
What does the feature vector project onto the weight space?
Only the POSITIVE projection of z onto the w-space
What are the 7 steps that a NN layer performs
Rotation, Reflection, Scaling, Rotation, Reflection, Bias, Positive Part
1989
Yann Lecunn Introduces CNNs which perform dimensional expansion without blowing up the number of weights, allowing us to compute the feature vector from an image
Three Properties of a Wave/Signal
Stationarity - Fluctuation of a Wave Does not Change Much - You will see similar patterns
Locality - Things close in domain-space are more likely to be correlated
Compositionality - The overall information is transmitted through the integration of the parts
How is stationarity represented in CNNs?
Parameter Sharing
1997
Jordan Schmidhuber introduces the long term short memory cell, a dynamical system (RNN) that is able to model long term dependencies
If you have multiple partial derivatives across each time step for the same weights, how do we resolve this in an RNN?
We sum the partials