Lecture 10: Operant Conditioning - Elements & Schedules Flashcards

1
Q

What is the fundamental component of operant behaviour?

A
  • Response is controlled by its consequences
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What are the 3 elements of operant behaviour?

A
  • Instrumental response
  • Outcome of the response (the reinforcer)
  • Relation/contingency b/n response and outcome (i.e. how they’re linked)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What does reinforcement usually lead to?

A
  • Increased uniformity of response pattern
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is stereotypy?

A
  • Uniformity of behaviour

- Short repeated segments of behaviour (eventually occurs b/c of brain mechanism)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

How can response variability be increased?

A
  • If the instrumental reinforcement procedure requires variable behaviour
  • “reinforcement of variability”, ex. L-R alternations
  • In the absence of explicit reinforcement of variability, responding becomes increasingly stereotyped with continued conditioning (ex. drawing triangles)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

How is the belongingness of the response relevant to operant conditioning?

A
  • Response - Reinforcer
  • Certain responses naturally ‘belong with’ the reinforcer b/c of the animal’s evolutionary history
  • Ex. Puzzle boxes: did not use yawning/scratching as they are not naturally linked to release, but operating latch/pulling chain are manipulatory responses naturally related to release
  • What makes sense to the animals?
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is instinctive drift?

A
  • Extra, interfering responses that develop during reinforcement training that are naturally-occurring behaviours related to the reinforcer
  • Ex. food reinforcer can push toward the development/interference of food-related responses
  • Can be quite strong and interfere with responses required by training procedures
  • Ex. raccoons will not put coin in slot, as instinct is to hold on
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is the Behaviour Systems Theory?

A
  • When a particular system is activated, associated behaviours are engaged
  • Includes both reinforced behaviours and those that are associated with instinctive drift (opposing motivations)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What happens when hamsters must dig or face-wash in order to receive a food reward? Why?

A
  • Digging is much easier to reinforce, as it is natural to behaviour system
  • Responses that become more likely when the animal is hungry are more readily reinforced with food
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Is the quantity of the reinforcer relevant in operant conditioning?

A
  • Yes
  • Rats run faster to receive more of the same reinforcer
  • How much effort is it worth?
  • Motivation is a factor
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Is the quality of the reinforcer relevant in operant conditioning?

A
  • Yes

- Dogs run faster to get more palatable reinforcers (sausage > dry food)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What happens when the quality of the reinforcer is shifted in operant conditioning?

A
  • Response changes
  • Ex. When a dog is expecting sausage and gets dry food, responding decreases drastically
  • Response increases again when the dog is again given sausage
  • Acting on expectation of previous outcome (learning history)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What did Crespi’s experiment find?

A
  • Responding to a particular reward also depends on animal’s past experiences with other reinforcers
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What is negative behavioural contrast?

A
  • When reinforcer decreases in previously-experienced quantity, response speed for reinforcer decreases
  • Ex. dropping number of pellets given => drop in response speed below original lowest responding
  • What you’re used to affects the change in behaviour
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What is positive behavioural contrast?

A
  • When reinforcer increases in previously-experienced quantity/quality, response speed for reinforcer increases
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What is response-reinforcer contiguity?

A
  • Delivery of reinforcer immediately after response

- A form of temporal relation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

What is response-reinforcer contingency?

A
  • Extent to which response is necessary and sufficient for occurrence of reinforcer
  • Relates to causal relation
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

How are temporal and causal factors related to each other in operant conditioning?

A
  • They are independent of each other
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

What effect on instrumental learning does delay of reinforcement have?

A
  • Instrumental learning is disrupted by delaying the reinforcer after the response
  • Ex. Occurrence of response/delivery of reinforcer was delayed
  • Increase in delay decreases response rate
  • With increasing delays, difficult to assign causality
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

Which response is ‘credited’ with reinforcement after delay?

A
  • Both temporal and causal factors are present

- Response-reinforcer relation requires temporal contiguity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

What is accidental/adventitious reinforcement?

A
  • Instrumental learning can develop as a result of this
  • Ex. No contingency, pigeon had food delivered every X seconds; superstitious behaviour emerged; claimed temporal contiguity drove learning
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

What are terminal responses?

A
  • Occurred more often at end of food-food interval

- Ex. Pigeons orienting to food magazine, pecking at magazine

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

What are interim responses?

A
  • Occurred more often near the middle of the food-food interval
  • Ex. Pigeons pecking at floor
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
24
Q

What influences the acquisition of instrumental behaviour?

A
  • Contiguity

- Contingency

25
Q

What are ratio schedules?

A
  • A response is reinforced only once a certain number of responses is made
26
Q

What are interval schedules?

A
  • A response is reinforced only if it occurs after a certain amount of time has passed
27
Q

What is a fixed schedule?

A
  • Same number of responses/specific amount of time
28
Q

What is a variable schedule?

A
  • Number of responses/time varies
29
Q

What is a fixed ratio schedule?

A
  • Fixed number of responses lead to reinforcement (ex. FR10)

- Steady and moderate rate of responding with brief, predictable pauses

30
Q

What are post-reinforcement pauses, and in what schedule are they found?

A
  • Zero responding just after reinforcement on FR

- Fixed Ratio

31
Q

What is the ratio run and in what schedule of reinforcement is it found?

A
  • High and steady responding that completes each ratio requirement
  • Fixed Ratio
32
Q

What is ratio strain and in what schedule of reinforcement is it found?

A
  • Zero responding after ratio requirement is increased too quickly (e.g. FR1 is switched to FR50)
  • Shock after experience
  • Switch gradually to prevent this
33
Q

What is a cumulative record?

A
  • Way to represent how a response is repeated over time

- Shows the total number of responses that occurred up to a particular point in time

34
Q

What is the simplest schedule of reinforcement?

A
  • Fixed Ratio 1

- AKA continuous reinforcement

35
Q

What is the Variable Ratio schedule of reinforcement? What is the rate of responding like?

A
  • Number of responses to obtain reinforcer varies from one reinforcement to the next (e.g. 10 responses for 1st, 13 for 2nd, 7 for 3rd, etc)
  • The average number of responses per reinforcer is what the VR schedule is labelled (e.g. VR10)
  • Steady rate of responding because response number is less predictable
  • Ex. scratch tickets, lottery
36
Q

What is a progressive ratio schedule of reinforcement?

A
  • Gradual increase in ratio

- A type of variable ratio schedule

37
Q

What is the fixed-interval schedule?

A
  • Amount of time that has to pass is constant from trial to trial (e.g. FI15)
  • Predictable
  • Performance reflects accuracy in timing (improves with training)
  • Low levels of responding
38
Q

What is the fixed-interval scallop?

A
  • Slower response rate immediately after reinforcement (NOT the same as FR post-reinforcement pause)
  • Gradually increasing rate across interval
39
Q

What is the variable-interval schedule?

A
  • Amount of time that has to pass is not constant from trial to trial, but varies, settling on an average for the session
  • Will define range
  • Maintain steady and stable rates of responding without pauses, similar to VR schedules
  • Slower than VR
40
Q

When does a pause in responding happen?

A
  • With a fixed schedule, where there is predictability
41
Q

What does the steepness of the cumulative record mean?

A
  • Rate of responding
42
Q

What schedule of reinforcement leads to the highest rate of responding?

A
  • Variable Ratio
43
Q

Which schedule of reinforcement leads to the lowest rate of responding?

A
  • Fixed interval
44
Q

What is another term for omission-training procedures?

A
  • Differential reinforcement of other behaviour (DRO)
  • Highlights that individual receives appetitive stimulus if they are engaged in behaviour other than response specified by procedure
45
Q

Who studied behavioural contrast effects?

A
  • Crespi
46
Q

What is temporal contiguity?

A
  • Delivery of reinforcer immediately after response
47
Q

What is response-reinforcer contingency?

A
  • Causal relation

- Extent to which instrumental response is necessary and sufficient to produce reinforcer

48
Q

What is a marking procedure?

A
  • Facilitates learning with delayed reinforcement
  • Mark target instrumental response in some way to make it distinguishable from other activities of the organism
  • Can be accomplished by introducing brief light/noise after target response, or moving animal to holding box for delay interval
49
Q

What is the learned-helplessness effect?

A
  • Effects of exposure to uncontrollable shock on subsequent escape-avoidance learning in dogs
  • Found that exposure to uncontrollable shock disrupted subsequent learning
50
Q

What is the triadic design used in studies of learned-helplessness effect?

A

Different exposures = different results

  • Escapable shock => rapid-avoidance learning
  • Yoked inescapable shock => slow-avoidance learning
  • Restricted to apparatus => rapid-avoidance learning
51
Q

What is the learned-helplessness hypothesis?

A
  • Assumes that during exposure to uncontrollable shocks, animals learn that shocks are independent of behaviour
  • Come to expect that reinforcers will continue to be independent of behaviour in the future
  • Undermines ability to learn a new instrumental response
  • Expectation of lack of control reduces motivation to perform instrumental response and even if they make the response and get reinforced, the previously learned expectation of lack of control makes it more difficult to learn that new behaviour is effective
52
Q

What is the difference between the learned-helplessness hypothesis and effect?

A
  • Effect = pattern of results obtained with triadic design (disruption of instrumental conditioning caused by prior exposure to inescapable shock)
  • Hypothesis = explanation/interpretation of the effect
53
Q

What is the attention deficit hypothesis?

A
  • Exposure to inescapable shock reduces extent to which animals pay attention to own behaviour (=> learning deficit)
54
Q

What are shock-cessation feedback cues?

A
  • Some of response-produced stimuli are experienced at start of escape response, just before shock is turned off
55
Q

What are safety-signal feedback cues?

A
  • Other response-produced stimuli are experienced as the animal completes the response, just after shock has been turned off at start of intertrial interval
56
Q

What is a schedule of reinforcement?

A
  • Program or rule that determines which occurrence of a response is followed by the reinforcer
57
Q

What is partial/intermittent reinforcement?

A
  • Situations in which responding is reinforced only some of the time
58
Q

What is the length of the post-reinforcement pause controlled by?

A
  • Upcoming ratio requirement
59
Q

What is limited hold?

A
  • Restriction on how long a reinforcer remains available

- Added to FI or VI schedules