Task 4 - M&M Flashcards
How does competitive learning work ?
-How does it find a categorisation?
- is unsupervised
- network finds a categorisation for itself based on the similarity between input patterns and the number of output units available
What do competitive learning networks learn?
learn to categorise input patterns into related sets, with one output unit firing for each set
What happens when an input pattern is presented to a competitive learning network?
output units compete with each other to determine which has the largest response
What happens with the connections to active and inactive input units to output units in competitive learning?
connections to winning output unit are strengthened and those from input units which were inactive are weakened
How are the weights set in competitive learning?
weights are set by prior learning of network, not by an explicit external teacher
What are the 3 phases of competitive learning?
excitation, competition and weight adjustment
To which connections only are weight adjustments done in competitive learning?
only made to connections feeding into the winning output unit
Which kind of learning rule does competitive learning use?
Uses a local learning rule
In which way are competitive networks a feature of many real brain circuits ?
- Can remove redundancy-> allocates a single output neuron to represent a set of inputs which co-occur
- they can produce outputs for different input patterns which are less correlated with each other than the inputs were
Which kind of associator is the auto- associator?
Form of pattern associator
What is the aim of the auto- associator?
to reproduce the same pattern at output that was present at input
What is the difference between pattern associator and autoassociator?
Auto-associator: output line of each unit is connected back to the dendrites of the other units -> recurrent connections
What is the netinput in an autoassociator?
Netinput: external input and internal input, generated by feedback from other units within the autoassociator
What are two features of auto-associators?
- Pattern completion
- Noise resistance
What does the pattern associator learn?
learns to associate one stimulus with the other
How does the training work for a pattern associator?
-What happens if learning is successful?
- training: pairs of patterns presented
- if learning successful: will recall one of the patterns at output when the other is presented at input
What is the pattern associator able to do after training?
after training: can also respond to novel inputs, generalising from its experience with similar patterns
How does pattern association take place?
pattern association takes place by modifying the strength of the connections between input units and output units
What are properties of pattern associators? (6)
- Generalisation: they generalise during recall
- Fault tolerance: graceful degradation; small damage still leads to correct response
- The importance of distributed representations for pattern associators: activity of all elements is used to encode a particular stimulus; generalisation and graceful degradation are only achieved if representations are distributed
- Prototype extraction and noise removal: recognition of a prototype that has never been seen
- Speed: fast due to parallel processing
- Interference is not necessarily a bad thing: allows generalisation, noise reduction and prototype extraction; one reason why it is tolerated is that the ability to generalize between stimuli is more useful than 100% accurate memory of specific past events
What is a recurrent network?
output line of each unit is connected back to the dendrites of the other units
What is the delta rule about?
output unit can be corrected by increasing the weights of connections from units in the previous layer which provide a positive input to it and by decreasing the weights of the connections which provide a negative input
What does saturation mean in context of Hebbian Learning?
If we continue to train the networks, it’s not possible that we can endlessly learn
What is sparseness?
measure of the proportion of units which will be active in that area in response to an input
What does sparse input enable?
sparse input enables an auto-associator to store more memories (e.g. dentate gyrus)
What is a difference between the Delta and Hebbian learning rule?
- Delta rule: gradually smaller changes of the weights
- Hebb rule: only one learning trial may be required to establish the necessary connections
Is it possible with Hebb’s law to build a computer model?
No, Hebb’s law is incomplete and inadequate to build a computer model
What are the problems of Hebb’s law when it comes to using it as a computer model?
- does not specify how much the connection between neurons should increase
- does not specify how to compute the activity of the two neurons
- there is nothing that ever allows the connection strength to decrease
- does not specify the exact conditions under which the connections should strengthen
What does Neo-Hebbian Learning consist of in its simplest form? (dynamical differential equations)
two sets of dynamical differential equations, one governing the activity change of an arbitrary network at a given instant in time and the other governing the weight changes of an arbitrary connection in the network at any instant in time
What is an instar?
From the point of view of the neurode, it receives a large number of stimulus signals, coming from somewhere “outside” its boundaries
–> from its perspective, it is the center of an inwardly radiating collection of such signals
What is an outstar?
- neurode sends its single output signal to a large number of other neurodes in the network
- when envisioning those outgoing signals as being more or less evenly distributed around the neurode, you can imagine an outwardly radiating star of output signals moving out from the neurode
What is each neurode in terms of instars and outstars?
every neurode: both the center of an instar, receiving incoming stimuli from the outside and the center of an outstar, transmitting its output back to other instars or the outside world
What do you look at in Neo-Hebbian learning?
you only look at change in connection strength
What are problems of Neo- Hebbian learning?
- behavior of the outstar matches actual classical conditioning not in detail
- weights only increase in strength BUT a biological system cannot possibly increase without bonds
How does the connection strength change in Differential Hebbian Learning?
the connection strength changes according to the change (difference) in the receiving neurode’s activation and the change in the incoming stimulus signal
What are the problems of Differential Hebbian Learning?
- outstar exists only in a single, continuously changing moment of time called now -> only learn from stimuli that appear simultaneously BUT Pavlov’s experiment: learned more quickly when bell was rung before
- outstar: acquisition curve is linear; BUT in real life: S-shaped
What do you look at in Differential Hebbian Learning?
you look at difference in change between connections
What happens in Differential Hebbian Learning if either neurode has a constant activity level?
- if either neurode has a constant activity level, it has an activity change of zero -> no learning occurs
- the activity changes can be either positive or negative, the weight change can be either positive or negative
Which two equations can be found in the Drive Reinforcement Theory (=DRT)?
- activity equation: describes the activation level of each neurode in the network
- weight change equation: describes how the connection strengths change during learning
How can incoming signals contribute to the weighted sum in Drive Reinforcement Theory (=DRT)?
- each incoming signal must individually be greater than the threshold -> can contribute to the weighted sum
- an incoming signal less than or equal to the threshold value is considered a zero signal
How do synaptic junctions work in DRT?
each synaptic junction is predetermined to be either an excitatory (positively weighted) or an inhibitory (negatively weighted) junction (always remains either positive or negative)
Can weights be zero in DRT?
No