Chapter 6 - Schedules of Reinforcement and Choice Behaviour Flashcards

1
Q

What is a schedule of reinforcement?

A

It is the rule that determines how and when a response will be reinforced

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What are the two types of reinforcement rates?

A
  1. continuous reinforcement

2. partial reinforcement

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is continuous reinforcement?

A

Reinforcing every correct response

  • most efficient way to condition a new response
  • rare in real life
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is partial reinforcement?

A

Reinforcing some, but not all responses

- more effective at maintaining or increasing the rate of response (and produces long lasting responses)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is the consequence of having different schedules?

A

There are distinct rates and patterns of responses with varying degrees of resistance to extinction

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What are the two basic types of partial reinforcement schedules?

A
  1. ratio

2. interval

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What do ratio types require?

A

A certain number of responses to be made before one is reinforced

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What do interval types require?

A

A certain amount of time to elapse before a reinforcer is given

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What are the two basic categories within ratio and interval partial reinforcement schedules?

A
  1. fixed

2. variable

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Describe fixed ratio (FR) schedules

A

FR schedules

  • reinforcement is given if the subject completes a pre-set number of responses (ex. FR5 = after 5 responses, the subject is rewarded)
  • continuous reinforcement (CRF) is the same as FR1
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What do we observe with fixed ratio schedules?

A
  1. ratio run: when the subject exhibits the same response again and again - shows that learning has occurres
  2. post-reinforcement pause: taking a break in performing after reward has been given
  3. ratio strain: a pause during the ratio run, following a sudden, significant increase in ratio requirement (they stop performing because of an increase in requirement)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Describe variable ratio (VR) schedules

A
  • the number of responses required to get each reinforcer is not fixed (it is random around a particular number - usually the average)
  • numerical value of the ratio indicates the average number of responses required per reinforcer (ex. VR7 means that on average, 7 responses are required before a reinforcer is introduced)
  • the reinforcer is less predictable, therefore there is less likelihood of pausing between responses (due to the fact that the very next response could be reinforced)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

When are interval schedules reinforced?

A

Only if they occur after a certain amount of time has passed

- time is the factor, but a response is still required to obtain a reinforcer

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Describe a fixed interval (FI) schedule?

A
  • a response is only reinforced if a constant or fixed amount of time has elapsed from the previous delivery of a reinforcer (ex. FI5 = five minutes have to elapse)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What is a fixed interval scallop?

A

Time towards the end of the interval will increase the rate of responding
- more responses closer to the time and fewer immediately after

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Describe a variable interval (VI) schedule?

A
  • a response is reinforced only if it occurs more than a variable amount of time after the delivery of an earlier reinforcer
  • because VI are less predictable, just like VR, there is a steady rate of response with little break in between
17
Q

Which two reinforcement schedules have the strongest response rate?

A
  1. Fixed ratio
  2. variable ratio
    - the organism learns that it is causing the reward with its behaviour
18
Q

What is the distinguishing characteristic of a fixed interval schedule?

A

The scallops

19
Q

Which schedule produces the highest rate of responding, and is the most resistant to extinction?

A

variable ratio schedules, because the organism doesn’t know when the response will be reinforced

20
Q

What are the similarities between ratio and interval schedules?

A
  • in both FR and FI schedules there is a typical pause after each reinforcer, and an increased rate of response just before the delivery of the next reinforcer
  • in both VR and VI schedules the response rate is steady without predictable pauses
21
Q

What is the difference between ratio and interval schedules?

A

Different rate of response even when reinforcement frequency is similar

22
Q

Why would different schedules with similar rates of reinforcement produce different rates of behaviour?

A
  1. reinforcement of inter-response times (IRT):
    - the faster you respond, the more likely you are to receive reinforcement (probabilistic)
    - variable ratio schedules reinforce shorter IRTs
    - interval schedules favour long IRTs
  2. feedback function:
    - higher responding is reinforced more (only in ratio schedules)
23
Q

Define choice behaviour

A

The voluntary act of selecting or separating from two or more things that which is preferred

24
Q

What factors are our everyday choices influenced by?

A
  1. reinforcer: quality and quantity
  2. behaviour: type of response, schedule of reinforcement
  3. available alternatives
  4. delay in reinforcement
25
Q

What is a concurrent schedule of reinforcement?

A

2 schedules of reinforcement are in effect at the same time

  • the subject is free to switch between responses
  • allows for continuous measurement of choice because the organism is free to switch between options at any time
26
Q

Describe Hernstein’s Matching Law

A
  • can only be used for prediction of VI schedules only
  • the relative rate of responding is matched to the relative rate of reinforcement
  • accurate in predicting generalized rates of reinforcement
  • takes into consideration all possible outcomes
27
Q

What is the formula for Hernstein’s Matching Law?

A

BL/BL+BR = rL/rL+rR

or

BL/Br = rL/rR

  • the rate of reinforcement from option A (ex. 480/20 = 24)
  • the rate of reinforcement from option B (ex. 480/60 = 8)
  • add these two (ex. 32)
  • divide rate of reinforcement of option A (24) by the total number of reinforcers possible (32) = 0.75
  • do same for option B = 8/32 = 0.25
  • answers should be proportions
28
Q

What is the implication of Hernstein’s Matching Law?

A

Rate of a particular response does not depend on the rate of reinforcement of that response alone

  • an operant response must compete with all other behaviours for an individual’s time
  • therefore, it is impossible to predict how a reinforcer will affect a behaviour without taking context into account
29
Q

What are the theories of choice behaviour?

A

maximization: organisms distribute their actions among response alternatives in order to maximize the amount of reinforcers possible in that situation
1. molecular maximizing: the choice is made according to the option that is most likely to be reinforced immediately
2. molar maximizing: the choice is made so that the highest number of reinforcers are collected in the long run (maximizing over time)

30
Q

What is a concurrent-chain schedule of reinforcement?

A

choosing one option makes the other one unavailable

  • two stages:
    1. choice link (the choice)
    2. terminal link (decided by the choice link)
31
Q

Define self-control

A

Choosing a large delayed reward over an immediate small reward
- more difficult to choose a long term reward over a short term reward because of the delay

32
Q

With a direct choice, what is an animal likely to choose?

A

An immediate and small reward

33
Q

With a concurrent-chain procedure, what is an animal more likely to choose?

A

A delayed large reward

34
Q

What is the value discounting function?

A
  • tries to anticipate what the decision will be
  • value of the reinforcer (V) is directly related to reward magnitude (M) and inversely related to reward delay (D)
  • k is the discounting rate parameter that indicates how rapidly reward value declines as a function of delay
  • when the value of the small reward decreases beyond the wait time, you would be more likely to wait for the large reward
35
Q

What is the formula of the value discounting function?

A

V = M/(1+kD)

36
Q

How does self-control look for people who are addicted to heroin?

A
  • they show steeper value discounting functions

- they are more impulsive and would take the immediate small reward over a delayed large reward

37
Q

How can delay discounting and impulsivity affect daily life?

A
  • slower reward discounting: associated with higher grades
  • faster reward discounting: associated with unsafe sex
  • slower delay discounting with age (because we are less impulsive as we grow up)
    • all depend on k factor **
38
Q

How can we train someone to have more self-control?

A
  1. training with delayed reinforcement increases likelihood of choosing a larger delayed reward in the future
  2. pre-commitment: make the decision to choose a larger delayed alternative in advance in a manner that is difficult or impossible to change later on
39
Q

What 5 steps can we take to modify our behaviour?

A
  1. identify the target behaviour
  2. gather and record baseline data
  3. plan your behaviour modification program
  4. choose your reinforcers
  5. set the reinforcement conditions and begin recording your progress