topic 4 Flashcards
Thorndike’s law of effect
if a response, in the presence of a stimulus is followed by satisfying state of affairs, the association between stimulus and response and strengthened and if it is followed by an unsatisfying event it is weakened
operant learning
a change in behavior as a function of the consequence that follows it.
escape behavior
when operant behavior increases by removing an ongoing event or stimulus. ex. pressing a lever to stop an electric pulse
avoidance behavior
when operant behavior increases by preventing the onset of an event or stimulus. ex. pressing a lever to prevent electric shock
discrete trial procedures
instrumental response produced once per trial. Each trial ends with removal of the animal from the apparatus. Ex. mice in a maze (how long it takes for them to reach the end= 1 trial) and then there is a “reset”
free operant procedures
animal remains in apparatus and can make many responses. ex. operant boxes
cumulative record
based on old cumulative recorder device.
unconditional (primary) reinforcer
a reinforcer that acquired its properties as a function of species evolutionary history. ex. food, sex. water, sleep
conditional (secondary) reinforcer
otherwise neutral stimuli or events that have the ability to reinforce due to a contingent relationship with other , typically unconditional, reinforcers.
variables effecting reinforcement
immediacy, specific reinforcer used, task characteristics, contingency, contiguity
immediacy
a stimulus is more effective as a reinforcer when it is delivered immediately after the behavior
contingency
a stimulus is more effective as a reinforcer when it is delivered contingent of the behavior.
contiguity
nearness of events in time (temporal contiguity) or space (spatial contiguity) - high contiguity is often referred to as pairing, less contiguity between the operant response and the reinforcer, diminishes the effectiveness of the reinforcer.
motivating operations
establishing operations make a stimulus more effective as a reinforcer at a particular time
abolishing operations
make a stimulus less potent as a reinforcer at a particular time
reinforcer magnitude
generally a more intense stimulus is a more effective reinforcer . relation between size and effectiveness is not linear.
schedule of reinforcement
a rule describing the delivery of reinforcement. different schedules produce unique schedule effects ( particular pattern and rate of behavior over time). over the long term, effects are very predictable. occurs in numerous species
continuous reinforcement schedule (CRF)
behavior is reinforced each time it occurs. rate of behavior increase rapidly. useful when shaping a new behavior. rare in natural world
4 types of intermittent reinforcement schedules
Fixed ratio schedule
variable ratio schedule
fixed interval schedule
variable interval schedule
Fixed ratio schedule (FR)
behavior reinforced after a fixed number of times generate post reinforcement pause
post reinforcement pause (PRP)
a period of inactivity or no responding that occurs immediately after a reinforcement is given. pausing typically increases with ratio size and reinforcer magnitude.
variable ratio schedule (VR)
the number of responses needed varies each time. ratio requirement varies around an average. PRP’s are rare and very short . produces higher rates than a comparable fixed ratio
random ratio
schedule is controlled by a random number generator. produces similarly high rates of responding. type of ratio used in casino and video games
progressive ratio
ratio requirements move from small to large. PRPs increase with ratio size.
fixed interval schedule (FI)
behavior is reinforced when it occurs after a given period of time. Produces PRP’s. responding increases gradually. uncommon in natural environment
variable interval schedule (VI)
the timing of the response needed varies each time. interval varies around an average. PRPs are rare and short. steady rates of responding. not as high as Variable ratio schedule. common in natural environments.
competing contingencies
the individual has multiple options that may conflict or compete for their attention or actions. ex. watch you tube or study
premack principle
in nature, different behaviours have different probabilities of occurring. e.g. eating is high probability and lever pressing is low probability.
Low prob -> high prob= reinforces low prob
high prob -> low prob= does not reinforce high prob
testing Premack principle
- establish baseline responding for different behaviours
- instrumental conditioning procedure w/:
L-> H H-> L
controlling stimulus (S)
an antecedent. a stimulus that changes the probability of an operant behaviour
discriminative stimulus (S^D)
aka occasion setter. a type of antecedent. a stimulus or event that precedes an operant and sets the occasion for its reinforcement.
extinction stimulus (S^∆)
a type of antecedent. a stimulus or event that precedes an operant and sets the occasion for non-reinforcement
establishing operation
makes a stimulus more effective as a reinforcer at a particular time. e.g. deprivation ( food is more effective stimulus if the individual is deprived of food)
discriminative stimulus
a cue or signal that indicates that a specific behavior will be reinforced
abolishing operation
makes a stimulus less potent as a reinforcer at a particular time. e.g., satiation ( food is less effective as a stimulus if the individual is full)
discrimination
occurs when the presence (or absence) of stimuli is the occasion on which a response will be followed by a reinforcement. e.g., key pecking is only reinforced when the green light is on -> the green light IS the occasion when pecking is reinforced. eventually the bird will only peck when the green light is on.
stimulus control
a change in operant behavior that occurs when either S^D or S^∆ is presented.
ex. “ the light color is functioning as a discriminative stimulus” or “the light has acquired stimulus control over the pigeons behaviour”
discriminative Index
a measure of the stimulus control exerted by an S^D or S^∆ .
generalization
less precise control of stimulus compared to discrimination. it is obtained by training a wide array of settings/ stimuli
stimulus generalization
- process where once a CS has been established )it produces CR reliably), similar stimuli may also produce a CR
or - process where once an operant response occurs to one discriminative stimulus, it also occurs to other similar stimuli
stimulus discrimination
- process where we exhibit less pronounced CR to CS’s that differ from original CS. -> CR occurs to CS but not another stimulus
or - process where less responding occurs to stimuli that are different from the original trained stimulus. -> operant response to trained stimulus but not others
discrimination training
S^D-> B -> C
ex. large dog -> child says dog -> praise
pony -> child says dog -> no praise
outcome= child says dog when they see a large dog but not a pony
~discrimination between categories
generalization training
S^D ->B -> C
ex. any sized dog -> child says dog -> praise
outcome= child will say dog when they see any dog
~ generalize within category