Ch. 6 Flashcards
Operant behaviors are influenced by their___
Elicited behavior is a function of what (precedes/follows) it; operant behavior is a function of what (precedes/follows) it.
Another name for operant conditioning is ____ conditioning.
Operant conditioning is ___
Classically conditioned behaviors are said to be ___ by the stimulus, while operant behaviors are said to be __ by the organism.
- Consequences
- Proceeds, follows
- instrumental
- a type of learning in which the future frequency (or probability) of a behavior is affected by its consequences. Note that this is essentially a restatement of Thorndike’s law of effect.
- elicited, emitted
emitted is used to indicate that operant behavior appears to have a more voluntary, flexible quality to it compared to elicited behavior, which is generally more reflexive and automatic.
Does this mean that operant behavior actually is voluntary? Not necessarily. Insofar as such behavior comes to be controlled by the consequences that follow the behavior, it can be argued that the sense of voluntariness, or “freedom of choice,” that accompanies such behavior is merely an illusion
Thorndike’s Law of Effect
According to Thorndike’s law of effect, behaviors leading to a satisfying state of affairs are strengthened, or “stamped in,” while behaviors leading to an unsatisfying or annoying state of affairs are weakened, or “stamped out.”
Thus, the extent to which the consequences of a behavior are satisfying or annoying determine whether the behavior will be repeated.
Thorndike (1898)
- interested in animal intelligence and thought could be properly assessed only through systematic investigation.
- Thorndike was not suggesting that animals could not in some ways be intelligent, but rather that we should not accept anecdotes as fact, nor should we assume that animals behaving in a particular way are doing so for intelligent reasons.
- Argue for caution in interpreting animal behavior.
Based on his research with cats, Thorndike formulated his famous law of effect, which states that behaviors that lead to a(n) satisfying state of affairs are strengthened, while behaviors that lead to a(n) unsatisfying state of affairs are weakened.
According to Thorndike, behaviors that worked were stamped in, while behaviors that did not work were stamped out
Thorndike’s puzzle box with the cat
In a typical experiment, a hungry cat was enclosed in a puzzle box and a dish of food was placed outside the box.
To reach the food, the cat had to learn how to escape from the box by stepping on a treadle that opened a gate.
- First time the cat was placed in the puzzle box, several minutes passed before it accidentally stepped on the treadle and opened the gate.
- Over repeated trials, it learned to escape the box more quickly.
- There was, however, no sudden improvement in performance as would be expected if the cat had experienced a “flash of insight” about how to solve the problem.
- Rather, it seemed as though the response that worked (stepping on the treadle) was gradually strengthened, while responses that did not work (e.g., clawing at the gate, chewing on the cage) were gradually weakened
- Thorndike suspected that a similar process governed all learning, and on this basis he formulated his famous law of effect.
Skinner’s Selection by Consequences
- Believed that behavior could best be analyzed as though it were a reflex. He also realized, like Pavlov, that a scientific analysis of behavior required finding a procedure that yielded regular patterns of behavior.
- Without such regularity, which could be achieved only in a well-controlled environment, it would be difficult to discover the underlying principles of behavior.
- “skinner box.”
- operant” procedure because the rat freely responds with a particular behavior (like pressing a lever) for food, and it may do so at any rate.
- experimenter controls the contingencies within the operant chamber, but the animal is not forced to respond at a particular time.
- This contrasts with other procedures for studying animal learning, such as maze learning, in which the experimenter initiates each trial by placing the rat in the start box.
The Skinner box evolved out of Skinner’s quest for a procedure that would, among other things, yield __ patterns of behavior.
In the original version of the Skinner box, rats earn food by ___ a ___ in another version, pigeons earn a few seconds of access to food by ___ at an illuminated plastic disc known as a ___
- regular
- pressing, lever, pecking, response key
Skinner’s procedures are also known as __ ___ procedures in that the animal controls the rate at which it earns food.
Skinner originally thought all behavior could be explained in terms of ______, but he eventually decided that
this type of behavior could be distinguished from another, seemingly more voluntary type of behavior known as ___ behavior.
- free, operant
- reflexes (or respondent behavior), operant
With the evolution of the Skinner box,
Skinner’s beliefs about the nature of behavior also changed.
He abandoned the notion that all behavior could be analyzed in terms of reflexes and, along with other learning theorists, came to believe that behaviors can be conveniently divided into two categories.
- One category consists of involuntary, reflexive-type behaviors, which, as Pavlov demonstrated, can often be classically conditioned to occur in new situations.
— Skinner referred to such behavior as respondent behavior. - The other category, which Skinner called operant behavior, consists of behaviors that seem more voluntary in nature and are controlled by their consequences rather than by the stimuli that precede them.
— It was this type of behavior that Thorndike had studied in his puzzle box experiments and upon which he had based his law of effect. It was this type of behavior that most interested Skinner as well.
— He spent the rest of his life investigating the basic principles of operant conditioning and applying those principles to important aspects of human behavior
Skinner’s definition of operant conditioning differs from Thorndike’s law of effect in that it is (more/less) mentalistic.
Skinner, however, was dissatisfied with Thorndike’s mentalistic description of consequences ___
- Less
- As being either satisfying or annoying.
Satisfaction and annoyance are internal states inferred from the animal’s behavior.
Skinner avoided any speculation about what the animal (or person) might be thinking or feeling and reworded the law of effect to emphasize the effect of the consequence on the future probability of the behavior.
Skinner’s principle of operant conditioning bears a striking resemblance to Darwin’s evolutionary principle of natural selection.
According to the principle of natural selection, members of a species that inherit certain adaptive characteristics are more likely to survive and propagate, thereby passing that characteristic on to offspring.
Thus, over many generations, the frequency of those adaptive characteristics within the population will increase and become well established.
Similarly, according to the principle of operant conditioning, behaviors that lead to favorable outcomes are more likely to be repeated while those that do not lead to favorable outcomes are less likely to be repeated.
Thus, operant conditioning is sort of a mini-evolution of an organism’s behaviors, in which behaviors that are adaptive (lead to favorable outcomes) become more frequent while behaviors that are nonadaptive (do not lead to favorable outcomes) become less frequent.
Operant conditioning is similar to the principle of natural selection in that behaviors that are ___ tend to increase in
frequency, while behaviors that are ___ tend to decrease in frequency.
The difference is that operant conditioning deals with changes at the level of a(n) ___ while the principle of natural selection deals with changes at the level of a(n) ___
- adaptive, nonadaptive
- individual, species.
The process of operant conditioning involves the following three components:
(1) A ____ that produces a certain
consequence.
(2) a ____ future likelihood of the
that serves to either increase or decrease the response that preceded it, and
(3) a _____stimulus that precedes the
and signals that a certain consequence is now available.
(1) response
(e.g., lever pressing produces a food pellet)
(2) consequence
(e.g., the consequence of a food pellet increases the rat’s tendency to again press the lever)
(3) discriminative response
(e.g, a tone that signals to the rat that a lever press will now
produce food).
operant behavior
is a class of emitted responses that result in certain consequences; these consequences then affect the future frequency (or probability) of those responses.
- call operants
- the probability or frequency of the response is often referred to as the strength of the behavior (probability, frequency, and strength should be regarded as equivalent).
- Note that, for the sake of simplicity, the definitions and examples will focus on whether the consequence increases or decreases the future frequency of a behavior.
— However, there are obvious limitations in how far that can progress, such that a consequence sometimes serves to simply maintain responding at a certain frequency. The manner in which certain patterns of reinforcement maintain a certain frequency of responding: schedule of reinforcement
-
Operant behavior is usually defined as a(n) ___ of responses rather than a specific response.
Remember that the terms response and behavior are essentially equivalent; however, behaviorists tend to use the term ____ when referring to a specific
instance of a behavior.
- Class
For example, there are many ways a rat can press a lever for food: hard or soft, quick or slow, right paw or left paw. All of these responses are effective in depressing the lever and producing food; therefore, they all belong to the same class of responses known as “lever presses.”
— Defining operants in terms of classes has proven fruitful because it is easier to predict the occurrence of a class of responses than it is to predict the exact response that will be emitted at a particular point in time.
For example, it is easier to predict that a hungry rat will press a lever to obtain food than it is to predict exactly how it will press the lever on any particular occasion. - Response
reinforcers
are those consequences that strengthen
behavior; that is, they increase its frequency or increase its probability.
A reinforcer is an event that follows a behavior and increases the future frequency of that behavior.
Punishers
Punishers are those consequences that weaken a behaviour that is; they decrease its frequency or probability
A punisher is an event that follows a behaviour and decreases the future frequency of that behavior.
The terms reinforcement and punishment refer to
the process procedure whereby the future occurrence of a behavior is
strengthened or weakened by its consequences.
Ex: Strengthening a roommate’s tendency toward cleanliness by thanking them when they clean the bathroom is an example of a reinforcement, while the thanks itself is a reinforcer.
Ex: Eliminating a dog’s tendency to jump up on visitors by scolding it when it does so is an example of punishment, while the scolding itself is a punisher
Diagrams of operant conditioning procedures generally use the following symbols.
Reinforcers are usually given the symbol SR (which stands for reinforcing stimulus), and punishers are given the symbol SP (which stands for punishing stimulus).
Lever press (R) - Food pellet (SR)
The food pellet is a reinforcer because it follows the lever press and increases the future probability of the rat pressing the lever.
Tell a joke (R) - Person frowns (SP)
The frown is a punisher because it follows the joke, and the future probability of Jonathan’s joke telling decreases
Note that, from a behavior analysis perspective, it is technically incorrect to say that a person or animal has been reinforced or punished.
Rather, it is the behavior that has been reinforced or punished.
Only the behavior increases or decreases in frequency.
There is value, however, in emphasizing the effect of the consequence on behavior.
If you want a child to stop doing something, should you tell them that their behavior displeases you or that they displease you?
Similarly, when your roommate does something that bothers you, will it be more constructive to tell them that their behavior disturbs you or that they disturb you?
Is it easier for people to change their behavior or to change who they are?
reinforcers and punishers
Reinforcer and punisher both refer to the specific consequence used to strengthen or weaken a behavior.
Note, too, that reinforcers and punishers are formally defined entirely by their effect on behavior.
Thus, the safest bet is to define consequences as reinforcers and punishers in relation to their effect on behavior and not in relation to how pleasant or unpleasant they seem.
a teacher might yell at their students for being disruptive, and as a result the students become more (not less) disruptive. Although the teacher is trying to punish the disruptive behavior, the yelling is actually having the opposite effect. By definition, therefore, the yelling is a reinforcer because it is causing the disruptive behavior to increase in frequency (perhaps because disruptive students find that other students admire them if they upset the teacher).
Weakening a behavior through the withdrawal of reinforcement for that behavior is known as ___. In general, this is a ___ process than punishment.
- extinction
- slower