5 NDC Feedforward Rate Networks and Backpropagation Flashcards
(28 cards)
How does an INDIVIDUAL rate neuron model perform simple computations?
What does this computation depend on?
by transferring its input into its desired output
depends on weight and threshold
What is the purpose of a linear decision boundary in rate neuron models?
to categories inputs into TWO group - one group elicits strong response
What is the shape of the response curve in rate neuron model?
sigmoidal
What type of neuron is in the Leaky Integrate and Fire model?
Are the dynamics of this model linear, quadatic or cubic?
-a single point neuron
-linear dynamics
What is the rate in the rate neuron model represent?
graded activity / firing rate ??
What does tau represent in the sigmoidal response curve of the rate neuron model?
As tau changes, what happens to this curve?
-parameter that characterises time scale
-smaller tau -> steeper sigmoidal curve and vice versa
What does u represent in the sigmoidal response curve of the rate neuron model?
-strength of input neuron receives
What type of values go in a truth table?
binary: 1 and 0
In a linear rate neuron model with two inputs, what is the equation?
And also with a threshold?
y=x1w1 + x2w2
theta < x1w1 + x2w2
In a rate neuron model as you change operation from AND to OR, what happens to the threshold (theta)?
threshold hold decreases
What kind of equation is the decision boundary of rate neuron model?
What is the equation? what is m?
-linear (equation of a straight line)
-x2=mx1+c
m is gradient
When theta equals zero, where does the (linear) decision boundary go through?
Why?
theta =0, goes thru origin of axes
because theta represents the the y intercept c
What are the three different activation functions? and their curve shapes?
linear - I (linear)
threshold - squarery S
logistic sigmoid - S
When is the response of the rate neuron model greatest?
when the pattern of inputs matches the pattern of weights
In the logisitic activation function eqn., what is w3 equal to?
minus theta
What is supervised learning?
learning from example or from a teacher
What is the point of using stochastic gradient descent in the rate neuron model?
to iteratively adjust the weights in the direction which reduces the error
What is the equation for gradient descent?
What does it mean?
๐๐ซ๐ซ๐จ๐ซ[๐จ๐ฎ๐ญ๐ฉ๐ฎ๐ญ]=(targetโ๐จ๐ฎ๐ญ๐ฉ๐ฎ๐ญ)^2
to get the error of the output, find the difference between the desired/target output and the output we have now (and the square it)
What is the output of a rate neuron model ? y=
weighted sum of inputs
โwi xi
What is the learning rate in the gradient descent recipe? What does it do?
Why canโt it be too big or too small?
alpha: is the magnitude of the adjustment of the weights/ the size of the steps
big: overfit of weights an miss optimal low error points
small: can cause the model to take forever to adjust to the ideal weight
Chain Rule:
If we have a function A[B] of a function B[C],
finding the derivative of A[B[C]] with respect to C is given by what?
๐๐ด[๐ต[๐ถ]]/๐๐ถ = ๐๐ด/๐๐ตร๐๐ต/๐๐ถ
In the gradient descent graph, what are the axes?
y = error[wi] x=wi
As the learning rate (alpha) decreases what happens to the speed of the descent ?
gradient descent is faster (smaller steps)
What is the chain rule of the gradient descent?
derror[output[weight]] / dwi