5 NDC Feedforward Rate Networks and Backpropagation Flashcards
How does an INDIVIDUAL rate neuron model perform simple computations?
What does this computation depend on?
by transferring its input into its desired output
depends on weight and threshold
What is the purpose of a linear decision boundary in rate neuron models?
to categories inputs into TWO group - one group elicits strong response
What is the shape of the response curve in rate neuron model?
sigmoidal
What type of neuron is in the Leaky Integrate and Fire model?
Are the dynamics of this model linear, quadatic or cubic?
-a single point neuron
-linear dynamics
What is the rate in the rate neuron model represent?
graded activity / firing rate ??
What does tau represent in the sigmoidal response curve of the rate neuron model?
As tau changes, what happens to this curve?
-parameter that characterises time scale
-smaller tau -> steeper sigmoidal curve and vice versa
What does u represent in the sigmoidal response curve of the rate neuron model?
-strength of input neuron receives
What type of values go in a truth table?
binary: 1 and 0
In a linear rate neuron model with two inputs, what is the equation?
And also with a threshold?
y=x1w1 + x2w2
theta < x1w1 + x2w2
In a rate neuron model as you change operation from AND to OR, what happens to the threshold (theta)?
threshold hold decreases
What kind of equation is the decision boundary of rate neuron model?
What is the equation? what is m?
-linear (equation of a straight line)
-x2=mx1+c
m is gradient
When theta equals zero, where does the (linear) decision boundary go through?
Why?
theta =0, goes thru origin of axes
because theta represents the the y intercept c
What are the three different activation functions? and their curve shapes?
linear - I (linear)
threshold - squarery S
logistic sigmoid - S
When is the response of the rate neuron model greatest?
when the pattern of inputs matches the pattern of weights
In the logisitic activation function eqn., what is w3 equal to?
minus theta
What is supervised learning?
learning from example or from a teacher
What is the point of using stochastic gradient descent in the rate neuron model?
to iteratively adjust the weights in the direction which reduces the error
What is the equation for gradient descent?
What does it mean?
๐๐ซ๐ซ๐จ๐ซ[๐จ๐ฎ๐ญ๐ฉ๐ฎ๐ญ]=(targetโ๐จ๐ฎ๐ญ๐ฉ๐ฎ๐ญ)^2
to get the error of the output, find the difference between the desired/target output and the output we have now (and the square it)
What is the output of a rate neuron model ? y=
weighted sum of inputs
โwi xi
What is the learning rate in the gradient descent recipe? What does it do?
Why canโt it be too big or too small?
alpha: is the magnitude of the adjustment of the weights/ the size of the steps
big: overfit of weights an miss optimal low error points
small: can cause the model to take forever to adjust to the ideal weight
Chain Rule:
If we have a function A[B] of a function B[C],
finding the derivative of A[B[C]] with respect to C is given by what?
๐๐ด[๐ต[๐ถ]]/๐๐ถ = ๐๐ด/๐๐ตร๐๐ต/๐๐ถ
In the gradient descent graph, what are the axes?
y = error[wi] x=wi
As the learning rate (alpha) decreases what happens to the speed of the descent ?
gradient descent is faster (smaller steps)
What is the chain rule of the gradient descent?
derror[output[weight]] / dwi
What is the gradient descent recipe?
- Identify the variables you need to change; ๐_๐ for a neuron.
- Define a cost function that you need to minimise to get the desired output:
The squared error, the squared difference between the neuronโs actual output and desired output. - Use the chain rule to find the derivative of the cost function with respect to the variable being updated; ฮ๐ค_๐
The derivative tells you how much change in the error will result from a change in ๐ค_๐
Moving in the direction of the derivative will increase the error; move in the opposite direction.
Multiply the derivative by a learning rate to adjust the step size of the learning rule.
๐ค_๐โ๐ค_๐โ๐ผโ ฮ๐ค_๐
what is the truth table like for the XOR exclusive or computation?
neuron response (output) given when light or bell is on but not both at same time
What are hidden neurons?
neurons behind output neuron in feedforward networks
What sort of network do you use to solve the XOR problem?
What computation are the left and right hidden neurons doing?
-feedforward neural network
-left: AND
right: OR