Ch. 7 Flashcards

1
Q

schedule of reinforcement

A

Is the response requirement that must be met to obtain reinforcement.

In other words, a schedule indicates what exactly has to be done for the reinforcer to be delivered.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

continuous reinforcement schedule

A

is one in which each specified response is reinforced.

An FR 1 schedule of reinforcement can also be called a continuous reinforcement schedule.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

intermittent (or partial) reinforcement schedule

A

is one in which only some responses are reinforced.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Schedule effects

A

Schedule effects are the different effects on behavior produced by different response requirements.

These are the stable patterns of behavior that emerge once the organism has had sufficient exposure to the schedule.

Such stable patterns are known as steady state behaviors.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Fixed Ratio Schedules

A

On a fixed ratio (FR) schedule, reinforcement is contingent upon a fixed, predictable number of responses.

For example, on a fixed ratio 5 schedule, a rat has to press the lever five times to obtain a food pellet.

Generally produce a high rate of response along with a short pause following the attainment of each reinforcer.

This short pause is known as a post-reinforcement pause.

For example, a rat on an FR 25 schedule will rapidly emit 25 lever presses, munch down the food pellet it receives, and snoop around the chamber for a few seconds before rapidly emitting another 25 lever presses.

Note, too, that each pause is usually followed by a relatively quick return to a high rate of response.

Thus, the typical FR pattern is described as a “break-and-run” pattern, a short break followed by a steady run of responses.

In general, higher ratio requirements produce longer post-reinforcement pauses.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Schedules in which the reinforcer is easily obtained are said to be

A

very dense or rich, while schedules in which the reinforcer is difficult to obtain are said to be very lean.

Ex: An FR 12 schedule of reinforcement is denser than an FR 75 schedule.

In general, “stretching the ratio”— moving from a low ratio requirement (a dense schedule) to a high ratio requirement (a lean schedule) should be done gradually.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

ratio strain

A

which is a disruption in responding due to an overly demanding response requirement.

Occurs when you jump from a low ratio requirement to a high ratio requirement

Ratio strain is what most people would refer to as “burnout,”

behavior may become increasingly erratic, with respect to what they’re trying to do. Like studying

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

variable ratio (VR) schedule,

A

On a variable ratio (VR) schedule, reinforcement is contingent upon a varying, unpredictable number of responses.

VR schedules generally produce a high and steady rate of response, often with little or no post-reinforcement pause. unlikely to occur when the minimum response requirement in the schedule is very low

As with an FR schedule, an extremely lean VR schedule can result in
ratio strain

Variable ratio schedules help account for the persistence with which some people display certain maladaptive behaviors.(gambling)

ex: Some predatory behaviors, such as that shown by cheetahs, have a strong VR component in that only some attempts at chasing down prey are successful. In humans, only some acts of politeness receive an acknowledgment,

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Variable ratio schedules of reinforcement may also facilitate the development of an abusive or exploitative relationship.

A

At the start of a relationship, the individuals involved typically provide each other with an enormous amount of positive reinforcement (a very dense schedule).

This strengthens the relationship and increases each partner’s attraction to the other.

As the relationship progresses, such reinforcement naturally becomes somewhat more intermittent.

In some situations, however, this process becomes malignant, with one person (the taker) providing reinforcement on an extremely intermittent basis, and the other person (the Giver) working incredibly hard to obtain that reinforcement.

Because the process evolves gradually (a process of slowly “stretching the ratio”), the Giver may have little awareness of what is happening until the abusive pattern is well established.

What would motivate such an unbalanced process?

— One source of motivation is that the less often the Taker reinforces the Giver, the more attention (reinforcement) they receive from the Giver.

In other words, the Giver works so hard to get the Taker’s attention that they actually reinforce the very process of being largely ignored by that partner.

may be relationships in which the partners alternate the role of Giver and Taker.

The result may be a volatile relationship that both partners find exciting but that is constantly on the verge of collapse due to frequent periods in which each partner experiences “ratio strain.”

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

fixed interval (FI) schedule,

A

On a fixed interval (FI) schedule, reinforcement is contingent upon the first response after a fixed, predictable period of time.

FI schedules often produce a “scalloped” (upwardly curved) pattern of responding, consisting of a post-reinforcement pause followed by a gradually increasing rate of response as the interval draws to a close.

On a pure FI schedule, any response that occurs during the interval is irrelevant.

Ex: If I have just missed the bus when I get to the bus stop, I know that I have to wait 15 minutes for the next one to come along. Given that it is absolutely freezing out, I snuggle into my parka as best I can and grimly wait out the interval. Every once in a while, though, I emerge from my cocoon to take a quick glance down the street to see if the bus is coming. My behavior of looking for the bus is on a(n) FI 15- min (use the abbreviation) schedule of reinforcement.

I will probably engage in a few glances at the start of the interval, followed a graduallv increasing rate of glancing as time passes

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

variable interval (VI) schedule,

A

On a variable interval (VI) schedule, reinforcement is contingent upon the first response after a varying, unpredictable period of time.

VI schedules usually produce a moderate, steady rate of response, often with little or no post-reinforcement pause.

Because VI schedules produce steady, predictable response rates, they are often used to investigate other aspects of operant conditioning, such as those involving choice between alternative sources of reinforcement.

ex: The behavior of checking your cell phone for notifications is maintained on a VI schedule. Our phones provide all sorts of reinforcers, from simple entertainment to meaningful communication with other people, and those little notification alerts pop up every so often. Even if you’ve just checked your phone, to see if any texts have arrived or if someone has liked the photo you posted, you may feel compelled to check it again.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Comparing the Four Basic Schedules

A

In general, ratio schedules tend to produce a high rate of response. This is because the reinforcer in such schedules is entirely response contingent, meaning that the rapidity with which responses
are emitted does affect how soon the reinforcer is obtained.

On interval schedules, the reinforcer is largely time contingent, meaning that the rapidity with which responses are emitted has little effect on how soon the reinforcer is obtained.

In general, variable schedules produce little or no post-reinforcement pausing because such schedules often provide the possibility of relatively immediate reinforcement, even if one has just obtained a reinforcer.

In general, fixed schedules produce post-reinforcement pauses because obtaining one reinforcer means that the next reinforcer is necessarily quite distant.

FR
Response rate: high
Post-reinforcement pause: yes

VR
Response rate: high
Post-reinforcement pause: no

FI
Response rate: increasing
Post-reinforcement pause: yes

VI
Response rate: moderate
Post-reinforcement pause: no

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

fixed duration (FD) schedule

A

On a fixed duration (FD) schedule, the behavior must be performed continuously for a fixed, predictable period of time.

For example, the rat must run in the wheel for 60 seconds to earn one pellet of food (an FD 60-sec schedule).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

variable duration (VD) schedule,

A

On a variable duration (VD) schedule, the behavior must be performed continuously for a varying, unpredictable period of time.

For example, the rat must run in the wheel for an average of 60 seconds to earn one pellet of food, with the required time varying between 1 second and 120 seconds on any particular trial (a VD 60-sec schedule).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

How do FD and VD schedules differ from FI and VI schedules?

A

On a VD schedule, reinforcement is contingent upon responding continuously for a varying period of time; on an FI schedule, reinforcement is contingent upon the first response after a fixed period of time.

Although duration schedules are sometimes useful in modifying certain human behaviors, such as studying, they are in some ways rather imprecise compared to the four basic schedules.

With FR schedules, for example, one knows precisely what was done to achieve the reinforcer, namelv, a certain number of responses.

On an FD schedule, however, what constitutes “continuous performance of behavior” during the interval could vary widely.</sub

—Ex: Julie’s son might read only a few pages during his two-hour study session or charge through several chapters; in either case, he would receive the reinforcer of being allowed to watch television.
—Remember reinforcing the mere performance of an activity with no regard to level of performance might undermine a person’s intrinsic interest in that activity. This danger obviously applies to duration schedules; one therefore needs to be cautious in their use.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

response-rate schedule,

A

in a response-rate schedule, reinforcement is directly contingent upon the organism’s rate of response.

three types of response-rate schedules.
— differential reinforcement of high rates (DRH)
—differential reinforcement of low rates (DRL)
—differential reinforcement of paced responding (DRP)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

differential reinforcement of high rates (DRH),

A

In differential reinforcement of high rates (DRH), reinforcement is contingent upon emitting at least a certain number of responses in a certain period of time or, more generally, reinforcement is provided for responding at a fast rate.

The term differential reinforcement means simply that one type of response is reinforced while another is not.

reinforcement is provided for a high rate of response and not for a low rate.

18
Q

differential reinforcement of low rates (DRL),

A

In differential reinforcement of low rates (DRL), a minimum amount of time must pass between each response before the reinforcer will be delivered or, more generally, reinforcement is provided for responding at a slow rate.

responding during the interval must not occur in order for a response following the interval to produce a reinforcer.

19
Q

differential reinforcement of paced responding (DRP),

A

In differential reinforcement of paced responding (DRP), reinforcement is contingent upon emitting a series of responses at a set rate or, more generally, reinforcement is provided for responding neither too fast nor too slow.

20
Q

noncontingent schedule of reinforcement

A

On a noncontingent schedule of reinforcement, the reinforcer is delivered independently of any response.

In other words, a response is not required for the reinforcer to be obtained.

Such schedules are also called response-independent schedules.

There are two types of noncontingent schedules: fixed time and variable time.

21
Q

fixed time (FT) schedule,

A

On a fixed time (FT) schedule, the reinforcer is delivered following a fixed, predictable period of time, regardless of the organism’s behavior.

22
Q

variable time (VT) schedule,

A

On a variable time (VT) schedule, the reinforcer is delivered following a varying, unpredictable period of time, regardless of the organism’s behavior.

23
Q

adjunctive behaviors.

A

When noncontingent reinforcement happens to follow a particular behavior, that behavior may increase in strength even though it is not required to obtain the reinforcer. Such behavior is referred to as superstitious behavior.

ritualistic behaviors observed may have consisted of innate tendencies, almost like fidgeting behaviors, that are often elicited during a period of waiting.

superstitious behavior thinking doing something will get you a reward when it doesn’t

24
Q

Herrnstein (1966)

A

noted that such behaviors can sometimes develop as a by-product
of contingent reinforcement for some other behavior.

25
Q

What happens if a noncontingent schedule of reinforcement is superimposed on a regular contingent schedule of reinforcement?

A

What if, for example, a pigeon responding on a VI schedule of food reinforcement also receives extra reinforcers for free?

the pigeon’s rate of response on the VI schedule will decrease.

26
Q

Noncontingent reinforcement benefits

A

noncontingent reinforcement is sometimes quite beneficial. More specifically, it can be an effective means for reducing the frequency of maladaptive behaviors.

Noncontingent reinforcement has even been shown to reduce the frequency of self-injurious behavior.

27
Q

Carl Rogers (1959)

A

the famous humanistic psychologist, called “unconditional positive regard.”
(noncontingent reinforcement)

Unconditional positive regard refers to the love, respect, and acceptance that one receives from significant others, regardless of one’s behavior.

Rogers assumed that such regard is a necessary precondition for the development of a healthy personality.

From a behavioral perspective, unconditional positive regard can be viewed as a form of noncontingent social reinforcement, which can indeed have beneficial effects.

In fact, it seems likely that proper child rearing requires healthy doses of both noncontingent reinforcement, which gives the child a secure base from which to explore the world and take risks, and contingent reinforcement, which helps shape the child’s behavior in appropriate ways, maximize skill development, and prevent the development of passivity.

28
Q

complex schedule

A

complex schedule consists of a combination of two or more simple schedules.

29
Q

Conjunctive Schedules

A

A conjunctive schedule is a type of complex schedule in which the requirements of two or more simple schedules must be met before a reinforcer is delivered.

For example, on a conjunctive FI 2-minute FR 100 schedule, reinforcement is contingent upon completing 100 lever presses plus at least one lever press following a two-minute interval.

30
Q

Adjusting Schedules

A

In an adjusting schedule, the response requirement changes as a function of the organism’s performance while responding for the previous reinforcer.

ex: To the extent that a gymnast is trying to improve their performance, their behavior is likelv on a(n) adjusting schedule of reinforcement; to the extent that their performance is judged according to both the form and quickness of their moves, their behavior is on a(n) conjunctive schedule of reinforcement.

31
Q

Chained Schedules

A

A chained schedule consists of a sequence of two or more simple schedules, each of which has its own SD and the last of which results in a terminal reinforcer.

In other words, the person or animal must work through a series of component schedules to obtain the sought-after reinforcer.

Within a chain, completion of each of the early links ends in a secondary reinforcer, which also functions as the
SD for the next link of the chain.

A chained schedule differs from a conjunctive schedule in that the two component schedules must be completed in a particular order, which is not required in a conjunctive schedule.

An efficient way to establish responding on a chained schedule is to train the final link first and the initial link last, a process known as backward chaining.

32
Q

goal gradient effect

A

The goal gradient effect is an increase in the strength and/or efficiency of responding as one draws near to the goal.

(Perhaps the most profound example of a goal gradient, however, is that shown by people who desperately need to urinate and move very quickly as they near the washroom.)

Responding tends to be weaker in the earlier chain.

33
Q

Chained Schedules and completing a course

A

Unfortunately, in the case of very long chains, such as completing a course, the terminal reinforcer is often extremely distant, with the result that behavior is easily disrupted during the early part of the chain (remember the goal gradient principle).

This is yet another reason for why it is much easier to be a diligent student the night before the midterm than during the first week of the semester.

Can anything be done to alleviate this problem?

One possibility is to make the completion of each link in the chain more salient (i.e., more noticeable), thereby enhancing its value as a secondary reinforcer.

Students sometimes keep detailed records of the number of hours studied or pages read.

They might also compile a “to-do” list of assignments and then cross off each item as it is completed.

Crossing off an item provides a clear record that a task has been accomplished and also functions as a secondary reinforcer that helps motivate us.

In fact, it can be so reinforcing that many people will actually add a task to a to-do list after they have completed it, simply to have the pleasure of crossing it off!

34
Q

backward chaining.

A

An efficient way to train a complex chain, especially in animals, is through backward chaining, in which the last link of the chain is trained first.

However, this type of procedure usually is not required with verbally proficient humans, with whom behavior chains can be quickly established through the use of instructions.

35
Q

drive reduction theory,

A

In drive reduction theory, an event is reinforcing to the extent that it is associated with a reduction in some type of physiological drive.

According to drive reduction theory, an event is reinforcing if it is associated with a reduction in some type of primary drive.

According to this theory, a secondary reinforcer is one that has been associated, directly or indirectly, with a primary reinforcer that reduces a drive.

A major problem with drive reduction theory is that some reinforcers do not seem to be associated with a physiological need.

ex: Thus, food deprivation produces a “hunger drive,” which propels the animal to seek out food. When food is obtained, the hunger drive is reduced. At the same time, the behavior that preceded this drive reduction, and led to the food, is automatically strengthened.

most theorists no longer believe that drive reduction theory can offer a
comprehensive account of reinforcement.

36
Q

Incentive motivation

A

Incentive motivation is motivation that is derived from some property of the reinforcer, as opposed to an internal drive state.

Ex: Research has shown that hungry rats will perform more effectively in a T-maze when the reinforcer for a correct response (right turn versus left turn) consists of several small pellets as opposed to one large pellet, The fact that several small bites of food are a more effective reinforcer than one large bite is consistent with the notion of incentive motivation

37
Q

The Premack Principle

A

The Premack principle proposes that reinforcers can often be viewed as behaviors rather than stimuli.

For example, rather than saying that
the rat’s lever pressing was reinforced with food, we could say that it was reinforced with eating food. It also distinguishes between behaviors in terms of their relative probability or frequency

The Premack principle states that a(n) high probability behavior can be used as a reinforcer for a(n) low probability behaviour

In the Premack principle of reinforcement, the low-probability behavior precedes the high-probability behavior, with the former being the target behavior and the latter being the reinforcer

ex: According to the Premack principle, if you crack your knuckles three times per hour and burp 20 times per hour, the opportunity to burp may be an effective reinforcer for knuckle cracking

Similar to the Premack principle, in Grandma’s rule, first you work, which is a(n) LPB, and then you play, which is a(n) HPB.

38
Q

Response Deprivation Hypothesis

A

The response deprivation hypithesis states that a behavior can serve as a reinforcer when

(1) access to that behavior is restricted and
(2) its frequency thereby falls, or is in danger of falling, below its preferred level of occurrence.

According to the response deprivation hypothesis, a response can serve as a reinforcer if free access to the response is restricted and its frequency then falls below its baseline preferred level of occurrence.

The preferred level of an activity is its baseline level of occurrence when the animal (or person) can freely engage in that activity.

The response deprivation hypothesis differs from the Premack principle in that we need only know the baseline frequency of the reinforcing behavior.
To help distinguish between the Premack principle and the response deprivation hypothesis, ask vourself whether the contingency is based on the frequency of one behavior relative to a second behavior (in which case, it is an example of the Premack principle), or the frequency of a single behavior which we are being, or are in danger of being, deprived of (in which case it is an example of the response deprivation hypothesis).

The response deprivation approach also provides a general explanation for why contingencies of reinforcement are effective. Contingencies of reinforcement are effective to the extent that they create a condition in which the organism is confronted with the possibility of a certain response falling below its preferred level.

39
Q

Behavioral Bliss Point Approach

A

According to the behavioral bliss point approach, an organism with free access to alternative activities will distribute its behavior in such a way as to maximize overall reinforcement.

Contingencies of reinforcement often disrupt the preferred distribution of behavior such that it is impossible to obtain the optimal amount of reinforcement.
When activities are not freely available such as when the two activities are intertwined in a contingency of reinforcement—then the optimal distribution may become unattainable.

Given this state of affairs, the organism will redistribute its activities to get as near to the bliss point as possible.

Ex: a rat that can freely choose between running in a wheel and exploring a maze might spend two hours per day running in a wheel and two hours exploring a maze— same amount of time exploring the maze as running. —> optimal distribution of activities for those two activities— that is, the behavioral bliss point for that particular rat.

40
Q

Kailey typically watches television for two hours per day and reads comic books for one hour per day.

You set up a contingency whereby Kailey must watch 2.5 hours of television each day in order to have any access to her comic books.

According to the Premack principle, this will likely be an ____ contingency.

According to the response deprivation hypothesis, this will likely be an ____ contingency.

A
  1. ineffective
  2. effective