Computational Neuroscience 2 Flashcards
Learning is achieved by…
…modifying the synaptic weights of the network
First learning algorithm that automatically changed the weights in neural networks came from…
…Widrow-Hoff trying to devise an automatic speech to text system
Widrow-Hoff rule
- Initial weights can have random, or zero, values
- A CS is presented and the output recorded
- The output is compared to a desired output, and if there is an error the weights are adjusted in order to reduce the error (supervised learning).
Same as the Rescorla-Wagner models.
Extinction trials
Trials in which the learned condition is ‘unlearned’. That is achieved by exposing the animal to the CS without the US.
The blocking effect
Pavlov suggested that it was only the co-occurrence of the CS and US that were key to conditioning.
However, by the late 60s several studies showed that in order to produce conditioning, an stimulus needs to have predictive power about the US.
Rabbits are separated in two groups: pre-trained group and sit-exposure group
Phase 1
• The pre-trained group is trained to the tone as CS before the airpuff US
• Meanwhile the sit-exposure animals sit in the experimental chamber.
Phase 2
• Both groups are trained to respond to a CS consisting of simultaneous tone and light (read the paper)
Error correction learning in the brain
The Rescorla-Wagner model requires a teaching signal to act as a feedback mechanism.
Therefore, it is not a model of learning at neuron level
However there is evidence that certain brain structure do provide a feedback signal that the system can used as a teaching signal.
• One example is the cerebellum — an area heavily involved in learning reflex responses
• In 1968, Richard Thompson suggested that there was a teaching signal in the cerebellum that the brain could use in conditioning
Error correction in the cerebellum
CS information reaches the cerebellum
Information about the US (airpuff)
travels through the inferior olive
Climbing fibers carry this information to the cerebellum
The output of the cerebellum is related to the conditioned response (eyeblink)
Part of the output also travels back to the inferior olive through inhibitory connections.
The pathway activated by the US acts as the desired output.
Thus the difference between the CR and the US determines the activity of the inferior olive
Inferior Olive activity = US - CR
Generalisation
The capacity to learn stimuli that have not been present in the training set, or the degree to which learning one stimulus is transferred to other stimuli.
Perceptron
An algorithm for supervised learning
Generalisation gradients in pigeons
Come back to
Generalisation in one layer network
Networks with local representation cannot generalise
Distributed representation
Coarse coding
It may allow the possibility of generalisation.
Representation is more robust; loss of a few neurons will not dramatically degrade the performance of the neural network.
In distributed representation, each stimulus is…
…mapped onto more than one node. The mapping can be topographic, but that’s not necessary for the network to perform. A key factor is that the degree of overlap between two stimuli represents their degree of similarity.