Lecture 9 - Adaptive Control Flashcards

1
Q

most ____ skills improve with practice

A

sensorimotor

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Proof that VOR learns to adjust

A

VOR restores retinal-image stability when vision is altered by lenses - magnifying glasses take ~ 30 min for VOR to grow stronger - minimizing glasses makes VOR grow weaker - glasses that flip the world upside down take about a week for VOR to adjust (VOR changes direction)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

what happens to saccades when eye muscle is damaged?

A

saccades miss target and drift

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

if eye muscle damage isn’t too sever, what happens to saccades?

A

neural adaption restore saccade accuracy and eliminate drift, even if the muscle is damaged

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

controller must know plant well, but…

A

we do not have to be born with accurate knowledge of our plant - plant changes through life - controller learns

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

How do control networks adjust themselves to improve performance?

A
  1. error-driven learning
  2. Learning by perturbation
  3. Gradient-descent learning
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

controllers learn the properties of their plants based on…

A

sensory feedback (learn from trying, examples…)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

what is the aim of learning?

A

minimize average error (aka risk, aka expected loss)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

error (e) =

A

y - y*

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

loss (L) =

A

|e|2/2

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What does the learner want?

A

minimize error & average loss E(L)

want both these to get 0

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

risk depends on…

A

probabilities of different situations (not every input has the same loss)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

how does the brain estimate risk?

A

Learning by perturbation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

learning by perturbation

A

make small changes to the weights and accept the ones that reduce error;

wi + η

If |epert| < |e|, then the perturbed weights are accepted

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

each decision to accept or reject perturbed weights is based on…but…

A

single input z, but overtime the neuron samples many inputs and the OVERALL risk is reduced

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What is the learning algorism that samples many weights and calculates the error every time?

A

weight perturbation

17
Q

What happens when η is too big? Too small?

A

big: learn faster, not accurate (can’t zero in)
small: learn slower, very accurate (steps are small)

18
Q

weight perturbation formula

A

wi ← wi + η(randomi - 0.5)

19
Q

Gradient-descent learning

A

find the minimum on a weight by weight plot

20
Q

equation for gradient-descent learning

A

Δw = -ηez

21
Q

What is Widrow-Hoff learning?

A

Gradient-descent learning

22
Q

Advantage of Widrow-Hoff learning

A

compute which way to go instead of guess (like perturbation learning does)

23
Q

Why is WH learning faster?

A

exploits knowledge of the network: neuron is linear and error is a linear function of the neruon’s out put signal

24
Q

learning rate of WH depends on …

A

η

too big: unstable, may overstep low ground

too small: slow

25
Q

how to learn nonlinear functions?

A
  1. weight perturbation (but too slow!!)
  2. error-backpropagation (generalized WH)
26
Q

backprop

A

gradient-descent algorithm that trains layer networks of non-linear cells;

each cell in each layer computers its own aL/aw and sense info upstream;

given appropriate signals from all its downstream cells, a neuron can compute its own aL/aw

27
Q

argument against backprop

A

takes precise communcation between layers to computer aL/aw for synapses deep in the network -> possibly not possible in the brain?

28
Q

what came about from backdrop being too complex?

A

shallow learning

29
Q

shallow learning

A
  • linear output cells
  • outer layer of neurons is adjustable
  • all other synapses are frozen (weights don’t change)
30
Q

curse of dimensionality

A

number of cells & synapses rise exponentially with dimension of the functions input space

31
Q

how to cope with curse of dimensionality?

A
  • have lots of synapses (seeds, spores, sperm, immune system, granuel cells in cerebellum)
  • prior knowledge