L16 - Neural Networks 2: Learning in ANNs Flashcards
Define supervised learning and unsupervised learning in context of a neural network. What’s the difference?
Supervised learning -> The output of the neural network is compared against known correct data. The accuracy of the model can then be determined and fine tuned.
Unsupervised Learning -> Used to establish patterns in data. No comparator output data. Output of the neural network is analysed to identify trends.
- The difference is that supervised is used mainly for prediction and classification purposes whereas unsupervised is used do identifying patterns in data.
What are the components of a Perceptron?
- A set of weighted connections that input data and associated weights into the perceptron.
- The activation function that operates the input data.
- The output axon, that outputs the function result.
What does the weights of each input data indicate?
The importance of that input
What type of learning is a perceptron for?
Supervised
Is a Perceptron an algorithm?
Yes. It’s a supervised learning algorithm mainly used for classification.
What does the threshold function do?
Gives a binary output
What does the threshold function not do?
Give any information regarding error.
What are the 4 steps of learning in the Perceptron?
- Initialise input weights and the threshold function.
- Present the input data as well as the expected output data.
- Calculate the output of the neural network…
1. Multiply each X by the corresponding weight W.
2. Sum all XW’s
3. Feed sum into activation function
4. Get output - Adjust weights based on amount of error between output and expected…
* Lots of error → Make big changes.
* Small error → Small incremental changes.
What are the formats of the weight update function?
- If the output Y is correct -> Wi(t+1) = Wi(t)
- If Y == 0, but should == 1 -> Wi(t+1) = Wi(t) + Xi(t)
- If Y == 1, but should == 0 -> Wi(t+1) = Wi(t) - Xi(t)
If the threshold function is OR, do we always floor the sum of the XW’s?
Yes… See letter notes for examples
See lecture notes for examples of perceptron learning with XOR and OR activation functions…
…
What parameter is used to control the learning rate of the weight update function?
n
E.g. Wi(t+1) = Wi(t) + nXi(t)
Define the Widrow-Hoff learning rule…
Weight updates of n are proportional to the error made.
What are some limitations of the single-layer perceptron? What is the solution to this?
- Can only solve linearly separable problems i.e Can’t solve XOR classification
- Multi-layer Perceptron
How does the multi-layer perceptron solve the issues of the single-layer perceptron?
- By adding hidden perceptron layers between input and output layer
- Activation function used provides information about error
- Error can be minimised as data is being passed through the algorithm