Operant Conditioning Flashcards
Edward L. Thorndike
Law of effect: positive consequences increase behavior; negative ones often have little or no effect
- His research on cats in puzzle boxes provided information about trial and error learning
** “predicts that behaviors that are followed by satisfying consequences will tend to occur again”
B.F. Skinner
** Skinner believed that bx is determined by the consequences that follow it and that, when a bx is reinforced, it increases.
Operant Conditioning
method of learning that employs rewards and punishments for bx
To answer these questions:
- First identify the target bx and determine if that bx is more or less likely to occur
- Then determine if the stimulus following the bx is being reinforced or punished.
Reinforcement:
Increases bx it follows
- Always involves an increase in bx
Punishment:
Decreases bx that precedes it
- Always involves decrease in bx
Positive
Application of a stimulus
“apply”
Negative
Withholding or removing stimulus
“withdraw”
Negative Reinforcement
an undesirable stimulus is removed to encourage the behavior
a graduate student smokes cigarettes in order to alleviate anxiety he feels about his ability to receive good grades in class”
- The student’s smoking continues b/c smoking removes (alleviates) anxiety
** “A person consistently avoids an object that produces a classically conditioned anxiety response because avoidance is negatively reinforced”
Positive reinforcement
A desirable stimulus is introduced to encourage certain behavior
the introduction of a desirable or pleasant stimulus (reward) after a behavior
Ex: Giving a child a treat when they were polite to their sister
Negative punishment
removal of a stimuli following a bx decreases the bx
Ex: charging telephone customers for using directory assistance to obtain phone numbers to decrease their use of directory assistance”
Positive punishment
introducing an undesirable stimulus (punishment) to discourage a specific behavior
Ex: spanking a child when rude to sister
Operant Extinction
gradual decrease in the response rate of a bx by the organism no longer being reinforced
Extinction burst
temporary increase in responding during extinction trials
Ex: rat has been reinforced for bar-pressing, sudden withdrawal of reinforcement will initially cause rat to press the bar more
Fading
gradual removal of a prompt
Behavioral contrast (aka positive behavioral contrast):
If reinforced for 2 dif bxs and reinforcement for 1 bx is withdrawn, the other bx is likely to increase
Reinforcers
a. Primary Reinforcer (unconditioned reinforcer): inherently desirable and do not depend on experience to acquire their reinforcing value (food, water)
b. Secondary Reinforcer (conditioned reinforcer): acquire their value only through repeated association w/ primary reinforcers (tokens, applause, gold stars)
c. Generalized secondary reinforcer: when a secondary reinforcer is paired with several different primary reinforcers (money)
Schedules of Reinforcement
- fixed interval
- variable interval
- fixed ratio
- variable ratio
Fixed interval
reinforcement is delivered after a fixed period of time regardless of the number of responses made (hourly, weekly wage)
- Produces low rates of responding b/c the number of responses is unrelated to the delivery of the reinforcement
Variable Interval
interval time b/w delivery of reinforcers varies in an unpredictable manner from interval to interval (“pop quizzes” throughout the semester)
- Produces steady but relatively low response rate
Fixed Ratio
reinforcer is delivered each time the subject makes a specific number of responses (piecework: workers receive payment after completion of a specific number of units)
- Produce relatively high, steady rate of responding
- “post-reinforcement response pauses”
Variable Ratio
reinforcers are provided after a variable number of responses (gambling)
- Produce the highest rates of responding and most resistant to extinction
Matching Law
organisms distribute their bx b/w two or more concurrent schedules of reinforcement.
- the frequency of responding to two or more alternatives (e.g., pressing two or more levers) matches the frequency of being reinforced for doing so.