PSYCH 1115 - Week 10 (ch7) Learning Flashcards
How do we learn?
When we learn, our minds naturally connect events that occur in a sequence.
This is called:
Associative Learning
(is constrained by biological predispositions - meaning if we vomit hours after eating a burger at a restaurant, we’ll develop an aversion to the food eaten, but not to the restaurant)
Classical Conditioning is when:
learning to link 2 stimuli.
Allows us to anticipate an event to which we have a reaction
EX: Pavlov discovered that salivation from eating food was eventually triggered by what should have been neutral stimuli such as seeing the food, seeing the food dish, seeing the serving person, just hearing the server’s footsteps coming with the bowl.
Neutral Stimulus: a bell (something that triggers no response)
Unconditioned Stimulus: food (something that naturally triggers a response without conditioning)
Unconditioned Response: dog salivates at food
———————————————-
Now take bell and use this device to start ‘Conditioning’:
Conditioned Stimulus: the bell
Conditioned Response: dog salivates
Then, can a 2nd device be used together with the bell to get a response? YES.
Called Higher-Order Conditioning (turning a neutral stimulus into a Conditioned Stimulus….by associating it with an already Conditioned Stimulus
Operant Conditioning is when:
Changing behaviour choices in response to consequences
RECALL:
Behaviourism:
behaviour can be explained in terms of conditioning, without appeal to the mind (thoughts, feelings).
Didn’t believe that the mental life was important to explore. Behaviourists believe that learning & conditioning are developed through associative learning (classical & operant), SO this means that anyone could learn any new task (doctor, lawyer, plumber, etc.) regardless of genetic backround, personality, or internal thoughts. All anyone needs is the right conditioning.
B.F. Skinner
John Watson
They didnt account for Cognitive Learning though - using observation & information, rather than direct experience.
But, how is learning defined?
refers to the relatively permanent change in a person’s behaviour to a given situation brought about by his repeated experiences in that situation, provided that the behaviour change can’t be explained on the basis of native response tendencies, maturation, or temporary states of the person or other animals.
THE 1ST STEP OF CLASSICAL CONDITIONING:
Define Acquisition:
The INITIAL stages of learning and conditioning.
How to know if acquisition has completed?
(how to know if there has been a successful association between a neutral stimulus and an unconditioned stimulus)
the unconditioned response (salivating) gets triggered by a conditioned stimulus (the bell – which used to be neutral)
WHEN to do this association?
Neutral Stimulus 0.5 sec before the Unconditioned Stimulus (bell before food always)
Define Extinction:
When the bell is still rung but no food is brought (stopped pairing CS + UC).
Diminishing of a conditioned response when the conditioned stimulus occurs repeatedly without the unconditioned stimulus.
Define Spontaneous Recovery:
a RETURN of the conditioned response (despite there being no further learning or relearning or conditioning done) after extinction occurred.
- Generalization:
- Discrimination:
- tendency to have conditioned responses triggered by related stimuli
ex: Pavlov conditioned dogs to drool when petted or scratched. - the learned ability to only respond to specific stimuli, to prevent generatlization
ex: Pavlov conditioned dogs to only drool at one specific bell with a specific pitch.
EXAMPLE
Systemic Desensitization:
Operant Conditioning:
Define Operant Conditioning:
A behaviour (a response) is followed by a reward/punishment
- Reinforced behaviour is likely to be done again
- Punished behaviour is less likely to be repeated
BF Skinner and his operant chamber.
In order for animals to start this operant conditioning, shaping of behaviour needed to be done based on 2 Principles.
- Works by building on natural behaviours (pigeons naturally walk around pecking things)
- Use method of successive approximations (as pigeons get closer and closer to the actual object that you want pecked, the pigeons get rewarded)
Thorndikes ‘Law of Effect’ essentially
Law of Effect?
An effect relating to Operant Conditioning
Discriminitive Stimuli:
the signal that indicates if a behaviour will be reinforced or punished
ex: pigeons will learn that they will only be rewarded if the screen turns blue (not red) – the monitor colour is the DS.
What are Reinforcers?
events or things that strengthen a response
- POSITIVE Reinforcer : add desirable stimulus
- NEGATIVE Reinforcer: remove aversive stimulus
ex: Oliver whines for treats (positive reinforcement)
Derek needs to get up from a nap to get him treats (negative reinforcement)
ex: Girlfriend keeps bugging BF to go dancing. He finally agrees. Next time she bugs, he immediately says yes (neg. reinforcement). Girl gets him to go dancing (pos. reinforcement)
ex: kid gets yelled at in classroom. Now he acts up even more in the classroom (positive reinforcement)
ex: kid doesn’t have to help parents wash dishes because earlier he helped to cook dinner
PRIMARY: meets basic needs, such as sex, food, cupcakes
SECONDARY: associated with primary reinforcers such as money or praise
(basically, primary satisfies biological needs, while secondary depend on learning)
-
Continuous Reinforcement: giving a rewards after the target every single time.
RESULTS: Subject gets desired behaviour quickly. -
Partial/Intermittent Reinforcement: giving rewards sometimes. Target behaviour takes longer to be acquired and established but persists longer without the reward.
RESULTS: usually results in steady rates of responding
Reinforcing Schedules:
- Fixed-Ratio Schedule: gets a treat every 3 times lever is hit, without fail. (ex: buy two drinks, get one free! or Maid takes 15 min break after cleaning 3 rooms)
- Variable-Ratio Schedule: gets a treat intermittently. Produces high rates of responding AND is difficult to extinguish. (ex: casinos - if the slot machine does pay out, I’ll just keep pulling the lever just in case! or if a charity makes an average of 10 calls before getting a donation)
BASED ON NUMBER OF BEHAVIOURS
- Fixed-Interval Schedule: a treat is rewarded every 2 minutes. As the treat time gets nearer, the rat will check the lever more frequently. (ex: baking cookies - will check the oven at 26 min, 27 min, 28 min, 28.5 min etc)
-
Variable-Interval Schedule : When a treat is rewarded at very random times throughout the day. Produces slow and steady responding
(ex: checking emails a couple times a day because I don’t know when the emails will come in)
BASED ON TIME-BASED INTERVALS