Final Deck Flashcards
Define stimuli
Anything that can be perceived by the animal
Animals are always learning- info is provided by everything in environment
Learned information then causes some kind of behavioral response
Types of stimuli
Unconditioned or primary stimuli- environmental information
Conditioned stimuli- cues/signals
What can be a cue
Anything the animal can perceive and trainer can use consistently
Define latency
time to respond to a cue
part of criteria- slow responses may have been shaped
How to condition the cue
Cue is associated with history of reinforcement:
Presence of cue- performance of a particular behavior is likely to be reinforced
Absence of cue- performance of a particular behavior is not likely to be reinforced
How to establish a cue
Shape behavior first
Teach cue after behavior is established
When you can predict animal will respond, start associating cue
Tips for training a cue
Give cue just as behavior is starting and gradually back up cue in time
Discrimination of cue and no cue
Work in blocks
Know cue is established once you have stimulus control
How do you know the animal understands the cue
Behavior always occurs immediately when cue is given
Behavior never occurs in absence of cue
Behavior never occurs in response to another cue
No other behavior occurs in response to the cue
How to transfer a cue
Pair second cue with presentation of first.
Commonly use a physical and verbal cue
Creates flexibility in how worked
Pros and cons of fading a cue
Cue can be made smaller in magnitude to make it less obvious to the audience. It can increase the animal’s attentiveness but is limited by the animal’s perceptual abilities
How to correct anticipation: animal offers behavior before it is asked
Correct by using set time frames
Set time goal for animal to NOT offer behavior
Reset time if animal offers behavior before goal is reached
Importance of feeding for attention/waiting
Define limited hold
Technique used to reduce latency when animal is not responding quickly enough to the cue
Reinforce for average response time, reduce over time with variability (like bell curve)
How to chain behaviors
Train last to first
Completion of one behavior cues start of next
Con- Animal may anticipate and skip steps in chain
Tips for trouble shooting behaviors
Discuss use of LRS
Think about application to ABCs of training
If something is not working, look at Antecedent and Consequence. Is there something in the cue/set up that is not allowing the animal to succeed? Is there something about the consequence not allowing the animal to succeed?
What is the goal of any session
To set the animal up to succeed
Why is the LRS beneficial
By using the LRS, we don’t reinforce incorrect behavior but allow animal to keep working
Pacing of session and handling LRS encourages good behavioral momentum
What determines whether something is reinforcing or not
Natural history
Subject’s state:
hungry, fearful, wanting attention, liking novelty
The arrival of reinforcement communicates what to the subject
The arrival of the reinforcer tells the subject what is being reinforced
(Baiting can reinforce a non-response)
(Negative reinforcer needs to cease as soon as desired behavior occurs)
Why is it beneficial to use the smallest size of reinforcement possible
Reduces time animal spends eating food
Spends more time learning
Takes longer to be full
Gives trainer more flexbility
Define jackpots
Reinforcement that is much bigger than normal
Surprise to the subject
May be used to reinforce a breakthrough
Define a continuous interval reinforcement schedule
AKA Continuous reinforcement (CRF)
With each behavior, a reinforcement is delivered. Every selected behavior is reinforced
Define a fixed schedule reinforcement schedule
When the responses are not reinforced every time they occur, but are reinforced at set times
Define Fixed Interval (FI)
the first behavior after a set period of time is reinforced. Ex: rat presses lever. This is limited to once a minute. So first press after that minute time frame is reinforced. Rat may keep pressing lever within the time window, but these are not reinforced
Define Fixed Ratio (FR)
Animal is reinforced after a set number of behaviors
So rat has to press lever 5 times to get reinforcement
Define Variable Interval (VI)
Time interval between reinforcements varies
Rat may be reinforced for touching lever once after a minute, then after 30 seconds, then after 5 minutes. Rat may keep pressing lever within the time window, but these are not reinforced.
Set a variance: average time and standard deviation time schedule moves around
Define Variable Ratio (VR)
Behavior reinforced after varying number of reiterations
Again need set variance with average and SD to vary around
Define differential reinforcement
AKA selective reinforcement
Adjusting reinforcement based on animals performance
Shaping tool
T or F: All behavior fall on a bell curve as to how “well” they are done
True. Using different schedules will change how much variability you have, and can be used to change how well an animal does a behavior
Define Random and Interrupted Reinforcement (RIR)
Reinforcers delivered intermittently in a randomized order.
So if a behavior occurred correctly, may or may not bridge, may or may not reinforce, and could deliver an unconditioned or conditioned reinforcement. All at random chance
Define Variable Ratio with Reinforcement Variety (VRRV)
Variable ratio schedule with the addition of secondary reinforcers in addition to primary ones
Define Continuous Reinforcement with Variable Ratio Primary Reinforcement Follow up (CRVRPRF)
With an animal that is well worked with, with a new behavior you always want to reward (ie continuous reinforcement), but you can deliver a variable primary reinforcer. Like using food, water, ice, etc. in varying amounts
As a new trainer, what type of reinforcement schedule should you use
Continuous schedule, fixed ratio. Also use when training a new behavior. With more experience (both you and your animal), introduce fixed ratio mixed with continuous, then can move to mix with variable ratio with differential reinforcement for established behaviors.
Use approximations to train reinforcement schedule
What type of reinforcement schedule will strengthen your behaviors
Variable schedules.
Why should you be careful about the timing of your reinforcement
You can accidently reinforce a negative behavior. Example: Polar bear pacing. Boomer ball given. Polar bear stops pacing. Ends up pacing more often because that was when reinforcement was given. Can also be positive- reinforcing while he is in water can encourage to spend more time in water and less time pacing
Define Least Reinforcing Stimuli (LRS)
Incorrect response leads to no change
A 3-5 second no response by the trainer for an incorrect response given by the animal following a signal.
Continue doing what you were doing- it is not changing the situation
What are the ABC’s of training
Antecedent
Behavior
Consequence