Lecture 7 - Variables that influence the effectiveness of conditioning and choice behavior Flashcards
What is a schedule of reinforcement?
A schedule of reinforcement is a program or rule that determines which occurrence of a response is followed by the reinforcer.
What is it called when every response results in delivery of the reinforcer?
continuous reinforcement (CRF)
What are the types of Simple Schedules of Intermittent Reinforcement?
Fixed-Ratio Schedule
Variable-Ratio Schedule
Fixed-Interval Schedule
Variable-Interval Schedule
Simple Schedules of Intermittent Reinforcement
What is the pattern of responding for Fixed-Ratio Schedule?
There is a steady and high rate of responding once the behavior gets under way. But there may be a pause before the start of the required number of responses
The zero rate of responding that typically occurs just after reinforcement on a fixed ratio schedule is called the post-reinforcement pause
The high and steady rate of responding that completes each ratio requirement is called the ratio run
What is a cumulative record?
A cumulative record is a special way of representing how a response is repeated over time
What is the post-reinforcement pause?
What is the ratio run?
The zero rate of responding that typically occurs just after reinforcement on a fixed ratio schedule is called the post-reinforcement pause.
The high and steady rate of responding that completes each ratio requirement is called the ratio run
What is ratio strain?
If the ratio requirement is increased a little (e.g., from FR 120 to FR 150), the rate of responding during the ratio run may remain the same. However, with higher ratio requirements, longer post-reinforcement pauses occur
This effect is called ratio strain
Simple Schedules of Intermittent Reinforcement
What is the pattern of responding for Variable-Ratio Schedule?
Because the number of responses required for reinforcement is not predictable, predictable pauses in the rate of responding are less likely with VR schedules than with FR schedules. Rather, organisms respond at a fairly steady rate on VR schedules.
Simple Schedules of Intermittent Reinforcement
What is the pattern of responding for Fixed-Interval Schedule?
As the time for the availability of the next reinforcer draws closer, the response rate increases. This increase in response rate is evident as an acceleration in the cumulative record toward the end of each fixed interval and is called the fixed-interval scallop.
Simple Schedules of Intermittent Reinforcement
What is the pattern of responding for Variable-Interval Schedule?
Like VR schedules, VI schedules maintain steady and stable rates of
responding without regular pauses
Interval Schedules
What is a limited hold?
A restriction on how long a reinforcer remains available
Comparison of Ratio and Interval Schedules
How does Reinforcement of Inter-Response Times result in different rate of responding?
If the participant is reinforced for a response that occurs shortly after the preceding one, then a short IRT is reinforced and short IRTs become more likely in the future. On the other hand, if the participant is reinforced for a response that ends a long IRT, then a long IRT is reinforced and long IRTs become more likely in the future.
Comparison of Ratio and Interval Schedules
What are feedback functions?
How do Feedback Functions result in different rate of responding?
the relationship between response rates and reinforcement rates calculated over an entire experimental session or an extended period of time
the faster the participant completes the ratio requirement, the faster it obtains the next reinforcer
Interval schedules have an upper limit on the number of reinforcers a participant can earn
Choice Behavior: Concurrent Schedules
Measures of Choice Behavior
How to calculate the relative rate of responding on each alternative?
BL/(BL+BR)
Choice Behavior: Concurrent Schedules
Measures of Choice Behavior
How to calculate the relative rate of reinforcement on each alternative?
rL/(rL+rR)
What is The Matching Law?
relative rates of responding match relative rates of reinforcement
BL/(BL+BR) = rL/(rL+rR)
BL/BR = rL/rR
What is the generalized form of the matching law?
How is under/overmatching and response bias accounted for?
BL/BR = b(rL/rR)^s
The exponent s represents sensitivity of the choice behavior to the relative rates of reinforcement for the response alternatives or the discriminability of the alternatives
undermatching - s < 1
b - response bias
What is the difference between molar theories and molecular theories?
Molar theories explain aggregates of responses. They deal with the distribution of responses and reinforcers in choice situations during an entire experimental session
In contrast to molar theories, other explanations of the matching relation operate on a shorter time frame and focus on what happens at the level of individual responses. Such explanations are called molecular theories
Mechanisms of the Matching Law
What is Molecular Maximizing?
organisms always choose whichever response alternative is most likely to be reinforced at a given moment in time
According to a more recent molecular account, a situation involving two response Alternatives A and B actually involves four different behavioral options: staying with Alternative A, switching from A to B, staying with Alternative B, and switching from B to A
What is Melioration?
operates on a time scale between molar and molecular mechanisms. Instead of aggregating data over an entire session or focusing on individual responses, melioration theory focuses on local rates of responding and reinforcement
Local rates are calculated only over the time period that a participant devotes to a particular choice alternative.
What is the concurrent-chain schedule of reinforcement?
To study how organisms make choices that involve commitment to one alternative or the other
The first is called the choice link. In this link, the participant is allowed to choose between two schedule alternatives by making one of two responses
The opportunity for reinforcement occurs only after the initial choice has been made and the pigeon has entered the terminal link
concurrent-chain schedules involve choice with commitment
the terminal link stimulus becomes a conditioned reinforcer
Self-Control Choice and Delay Discounting
What might the mathematical form of the value discounting function be?
The current consensus is that the value of a reinforcer (V) is directly related to reward magnitude (M) and inversely related to reward delay (D), according to the formula where k is the discounting rate parameter
V = M/(1+kD)
Self-Control Choice and Delay Discounting
What is the effect of longer delays from the small and large reward?
Higher tendency to select the larger reward
What can delay discounting be affected by?
Delay from the smaller and larger reward
Personality traits
Training
How can self-control in delay-discounting be trained?
- training with delayed reward
- a shaping procedure in which the large reward is initially presented without a delay, and the delay is then gradually increased across trials
- introduce a distracting task during the delay to the large reward or distract attention from the large reward during the delay period