Neural nets 2 Flashcards
What problems can SLPs not solve?
XOR problems (non linearly separable problems)
What is the difference between SLPs and multi-layered perceptrons?
MLPs have hidden layers between the input and output nodes.
What is the benefit of a hidden layer?
A hidden layer of neurons
can transform a non-linear
problem into a number of
simpler linearly separable
problems
Name a problem with training MLPs
Credit assignment problem. Which weights cause the problem? Hidden->output layer weights or input->hidden layer weights?
How do you train MLPs?
target (t) – output (y) error/cost term
What is backpropagation used for?
Backpropagation is used to solve the credit assignment problem. We want to back-chain
the error from the output layer to the input layer to update the weights of the network layer by layer
What are the 8 backpropagation algorithm steps?
- Present the data vector at the
input layer, - Pass weighted input layer
activations to hidden layer - Pass weighted hidden layer
activations to output layer - Apply the target value to the
output layer - Calculate the δ 𝑜𝑜 output error
values - Update the hidden-output layer
weights using using the δ 𝑜𝑜 values - For each hidden node, calculate its
δℎ error value using δ 𝑜𝑜. - Update the input-hidden layer
weights using the δℎ values.
Step 1-3 Forward Pass
Step 4-6Backward Pass
But how do we find the delta (error) values for each layer?
*Use gradient descent
principle
*Use the chain rule (from
calculus)
What is a problem with the gradient descent principle?
We can get stuck in a local minima. We would like to find the global minimum but with gradient descent we can’t be sure we have.
How do we solve the credit assignment problem?
1) Backpropagating (or backchaining) network error (and proxy error)
terms through the layers of weights in the network using calculus;
2) All nodes’ activation from the forward pass is maintained in order to appropriately assign error to the corresponding weights (if the
weights aren’t connected to active nodes they can’t have caused the error)