Recurrent Networks & Associative Memory Flashcards

1
Q

How does Pennartz broadly describe the advantages of the Hopfield model as a model of neural networks?

A

computationally tractable, yet very insightful

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Performatively what are the advantages of the Hopfield model? (5)

A

*Capable of handling noisy, fuzzy and/or incomplete information (“pattern completion”)

*Robust and fault tolerant; “graceful degradation” (when network loses cells)

*High learning capacity and flexibility; adapt to new environmental input without the need to be re-programmed

*Massive parallelism; many simultaneous operations

*High speed of computation (# logical operations/sec.)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

How does the architecture of a recurrent neural network differ to that of a CNN? What is it often used to model?

A

Often used for long-term associative memory or working memory

Recurrent neural network: all (or most) neurons project to each other:
* learning capacity distributed across recurrent (“associative”) connections
* connectivity can be “full” or partial / sparse

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is a prototypical case of these recurrent auto-associative memory networks?

A

Hopfield model of (auto)-associative memory

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is a requirement of the Hopfield model

A

Store a set of patterns in such a way that when presented with a new pattern P, the network responds by producing whichever one of the stored patterns most closely resembles P

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What are the chief features of a Hopfield model?

A
  • fully connected, recurrent network (but no autapses; self activation)
  • can do asynchronous parallel processing
  • content-addressable memory
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

How would an image look in the Hopfield network

A

Black and white images in which each picture pixel
is a neuron (black=inactive, white=active; all connected)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

How are these Hopfield networks related to Alzheimers?

A

Take away synapses and neurons –> Still able to recognise pattern –> related to Alzheimer, only a huge loss leads to visible changes or decline

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What do the train and test stages look like for a Hopfield network?

A

Neural networks are first trained on input patterns,
and then tested on a cognitive problem

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Describe the training phase

A

Training phase of auto-associative memory task:
* Input pattern is presented; tuned neurons are activated
* Distributed pattern of activity leads to distributed weight changes (Δwi,j)
* Each new trial (iteration) leads to an input pattern being ‘imprinted’ which is implemented by updating of synaptic weights (~LTP, LTD)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What does the test phase consist of?

A

Present test pattern: does network complete/retrieve the pattern?

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What could this imprinting be compared to in ecology?

A

imprinting in newborn animals:
Grouse chick imprints mother features
=> Chick recognises mother even when occluded/noisy/partially complete
(Use: memory & Gestalt grouping principles)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

In neurobiological networks, do neurons code “pixels”?

A

No, rather: features, present in a receptive field
Small RF, V1 simple cells; simple stimuli
Large RF, e.g. inferotemporal cortex; more complex stimuli, size invariance

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What activation function does the Hopfield network use?

A

ai=1 if Σwijaj > Threshold value (usually 0) (bi)
ai=0 if Σwijaj < Threshold value

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What learning rule does the Hopfield network utilise? Describe what we want to do conceptually

A

Learning rule / information storage algorithm:
We want to “imprint” a pattern p in the network, basically
according to Hebb’s rule:
* case 1: pre- and postsynaptic activity correlated (both ai= +1)
=> set their connecting synapse to a value of +1 (also: if both ai=0, aj= 0)

  • case 2: pre- and postsynaptic cell anti-correlated
    (only one neuron with a= +1, the other 0)
    => set their connecting synapse to a value of -1
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

How do we compute this rule mathematically?

A

Δwij = (2ai-1)(2aj-1):
Δwij = (2(1)-1)(2(1)-1) = 1
Δwij = (2(0)-1)(2(0)-1) = 1
Δwij = (2(1)-1)(2(0)-1) = -1

Next: store a set of N patterns (states) in the network:
Δwij = Σ(N_p=1) (2ap_i -1)(2ap_j-1)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

What happens if you store too many patterns relative to the number of neurons?

A

If you store too many patterns relative to the amount of neurons that you have you have a cancellation effect

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

Is all of this physiologically plausible?

A

Firstly separating learning from testing is a lil artificial and being fully connected is not super realistic

Secondly this model works in terms of LTP; both neurons excite each other and strengthen synapse. This doesn’t really make sense with LTD however, as activation is required for LTD so a=0 could not be the cause (?):

Change of the weight between pre to post neuron = modification of postsynaptic activity (+1)

Two neurons have activity of +1 = 1 product of equation is = 1

Two neurons have activity of 0 = 1 –> LTP
One neuron a = 1 and other neuron a = 0 –> -1 –> LTD

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

How does the network evolve from one state to the next?

A
  • We pick out two neurons, which were anti-correlated during “imprinting”
  • This implies that their reciprocal weights should be negative
  • Recall phase: during the “test” phase, no further weight changes occur
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

Describe a “tense situation” and how it is resolved in a recurrent network. Does this translate to biology?

A

“Tense situation”: activity states do not conform to pre-set weights. If two neurons both had an activation of 0 previously and therefore had a negative weight, however now down have an activation of one. We rely on existing weights that have been established during learning. One neuron will
win out over the other neuron (due to noise). Now the activity state is anticorrelated, according to the pre-set weights:
a ”relaxed situation”; Evolving to the attractor state

Biology e.g. Once one of the two neurons inhibits the other, the second can’t inhibit the first one anymore –> one neuron will win out over the other

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

What is meant by content-addressable memory in the context of these networks?

A

Stored content is not retrieved using a special ‘label’ (address) such as in Von Neumann computers, but using a part of the memory content itself

22
Q

What is the storage capacity of these networks?

A

0.15 N
(N=number of neurons)

23
Q

Why do these networks have a storage capacity?

A

If exceeding storage capacity –> things start influencing each other

24
Q

What are emergent collective properties of these networks?

A

*Memories are stable entities / “Gestalts”
*pattern completion & recognition
*generalization & categorization
*error correction, noise tolerance (Noisy image can be ‘cleaned up’)
*graceful degradation
*time sequence retention (not here)

25
Q

What is meat by time series retention?

A

If you have one pattern, its retrieval can lead to the next one; e.g picture of a spider to the next picture

26
Q

What does a stored pattern ‘represent’ in the model?

A

If you present an arbitrary test pattern to a network that
has already stored a set of different patterns, the network converges to a state corresponding to the most similar stored
pattern. Think of the energy landscape with a global maximum, and the ideal pattern as a global maximum.

A stored pattern thus represents a stable limit point or attractor

27
Q

What do the axes represent in the energy landscape?

A

Z-axis: amount of energy
X-Y plane: possible states that the system (network) can occupy

e.g. 3 units, activity 0 or 1:
2^3 = 8 possible states
squash cube with 8 positions
onto X-Y plane

28
Q

What analogy can you use for the state of the system?

A

Ball (network state) rolls down the slopes of the energy landscape until arriving at a global or local minimum (attractor state)

29
Q

After training, what is the test pattern?

A

Test pattern: Network states evolve according to the activation, the energy E always decreases (or does not change)

30
Q

Give the equation for the energy, what is the assumption here?

A

E = -1/2 sum(wij ai aj)

1/2 is bias so not a constant but the negative value is important

But, under the assumption: wij = wji (thus: symmetrical synapses)
This assumption is not realistic for neurons, but convenient for introducing Energy

31
Q

Give an example of a situation of high energy and low energy i regards to the weights and activations

A

Example situation of high energy / “tension”: wij= -1 but (aiaj)= 1
Example situation of low energy / “relaxation”: wij= -1 and (aiaj)= -1

Works with values of +1 and -1 as this is more easily proven

32
Q

When does ai = 1

A

if sum(wijaj) > 0

33
Q

What does Vi represent and how is it calculated?

A

Vi: total postsynaptic input
= sum(wijaj) – bi = sum(wijaj)

34
Q

What is the change of energy if wij = wji?

A

ΔE = -ΔVi . sum(wijVj)

35
Q

Why does the function decrease monotonically (moves or doesn’t) over iterations mathematically?

A

if ΔVi > 0 then sum_j(wij aj) > 0
if ΔVi < 0 then sum_j(wij aj) < 0

36
Q

So….energy is guaranteed to move downhill (or stay same). What problem remains?

A

System may get stuck into a local minimum, yielding a “spurious state”

e.g reversed states or “mixture states” e.g mix between two faces

37
Q

What are some psychophysical equivalents

A

ambiguous figures, binocular rivalry, illusions

38
Q

What is a common element of these physiological mixture states?

A

Sensory input is associated with different / conflicting sources of evidence:
* Memory / semantic meaning (rabbit-duck)
* Left and right eye (binocular rivalry)
* Gestalt features (local – global)

Beware! interpretation as “attractor” is being made, but still speculative

39
Q

How can we subsequently make the model more biological? (4)

A

*Get rid of symmetry in the synaptic weights
(wij = wji)

*go from full connectivity to sparse connectivity

*go from “iteration steps” to dynamics in real
time

*incorporate more neurobiological features, e.g.
excitatory and inhibitory cells, physiological
membrane properties

40
Q

In CA3 network of the hippocampus what is the connectivity? Is this enough for memory recall? Why?

A

CA3 network of hippocampus: recurrent connection probability is ~3%

Yes, you do lose capacity but also less chance of overlap –> things still work

41
Q

How was the computational capacities of the models expanded in subsequent work? (3)

A
  • Storage & retrieval of sequences (“movies”),
    as a more complete model of episodic memory
    (delay lines, D)
  • Continuous attractor models: string of stable
    points (e.g. place cells, grid cells)
  • “Unlearning” or active forgetting to stabilise
    memory storage
    Passive decay of potentiated synapses?
    Generally more useful: activity-dependent
    decrements of strength
    Theory-driven search for “long-term
    depression”
42
Q

What is the physiological equivalent to Hopfield’s learning rules?

A

They relate to long-term potentiation: a cellular model for learning- and memory processes

43
Q

Name two useful properties of LTP for memory

A

LTP has (at least) 2 very useful properties for memory:
-synapse specificity: If a connecting neuron is active, the synapse will be strengthened; If a neuron is inactive the synapse will not be strengthened

-associativity: If a connecting neuron has strong inactivation, the synapse will be strengthened; If a neuron has weak activation the synapse will also be strengthened

44
Q

What molecular structures are crucial for this associative plasticity?

A

NMDA receptors are crucial for this associative plasticity

45
Q

How does the model learning rule therefore compare to the physiology?

A

A pattern p is “imprinted” in the network:

case 1: pre- and postsynaptic activity correlated (both a = +1)
=> set their connecting synapse to a value of +1
This relates to LTP and corresponds with Hebbian learning

But also: if both ai=0, aj= 0 this also sets their connecting synapse to a value of +1
This is unrealistic
_______________________
case 2: pre- and postsynaptic cell anti-correlated
(only one neuron with ai= +1, the other 0)
=> set their connecting synapse to a value of -1

If aj= 1, ai=0 => ~ LTD; Arguments could be made here
If aj= 0, ai= 1 => no change; This works

46
Q

Describe subsequent work findings regarding this:
If aj= 1, ai=0 => ~ LTD

A

LTD and depotentiation:
*LTD can be induced by low-frequency stimulation
*Has input-specificity, like LTP
*Low-frequency also works to reverse LTP, partially
~ partly compatible with anti-correlation rule (Hopfield)

47
Q

Describe experimental setup of subsequent work regarding this model & Genomics and cognition

A

*NMDA receptor KO in hippocampal area CA3

*Pattern-completion task: retrieve full memory using incomplete sets of cues

*Motivation to choose CA3: relatively large amount of recurrent connections between pyramidal cells
Idea: NMDA receptors crucial for plasticity during “imprinting” phase

Normal learning in standard Morris Water maze
with 4 spatial cues between
Mutant (CA3-KO)
Floxed control;
Cre control mice
Had to remember a target quadrant

48
Q

What were the results of the study?

A

Complete pattern –> normal mice can find platform

Knockout NDMA receptor
If familiar with cues, knockout can find platform from complete pattern

Probe trial with partial cueing:
Degraded performance in mutants

49
Q

What conclusion can be drawn about importance of CA3 for pattern completion / retrieval?

A

-caveats: do CA3 mutants have intact visual capacities?

-do these mice navigate based on only 1 cue that remains (green triangle)?
(was this really a deficit in spatial learning, i.e. learning of spatial configurations?)

  • a single-cue task is going to be more difficult anyway (apart from pattern completion)
50
Q

Describe subsequent work investigating systems neurophysiology evidence that systems need time to converge to recognition state

A

*Natural object images are mixed with random noise patterns of equal spatial frequency & luminance

*Sample-delay-match / non-match test:
-Match case: release lever, get reward
-Non-match: hold on to lever, get reward later
-test image: non-degraded object

*Single-unit recording in Lateral PFC,
around principal sulcus