Learning Flashcards
Instinctual behaviours
Examples: imprinting, homing, migratory behaviours, etc.
Pavlov’s experiment 1890s
NS –> no response
US –> UR
Repeatedly pair NS and US –> UR
Results: CS –> CR
Reflexive behaviours
Examples: eye-blinking, ‘sucking’ and ‘gripping’ in babies; some reflexive behaviours may disappear as you grow older
Habituation
Decline in the tendency to respond to stimuli that have become familiar due to repeated exposure; ensures that benign stimuli do not interrupt our activity or cause us to expend unnecessary energy
Classical conditioning
A neutral stimulus is repeatedly paired with a stimulus that automatically elicits a particular response –> previously neutral stimulus becomes a conditioned stimulus that also elicits a similar response; classical conditioning is not so much the replacement of the US by the CS, but a learning mechanism where the CS (and the CR) prepare the animal for the onset of the US and UR
Watson & Raynor 1920: Little Albert experiment
Conditioned fear (fear associated with certain stimuli)
Fetishes
A person has heightened sexual arousal in the presence of certain inanimate objects, with the object becoming a conditioned stimulus that can elicit arousal on its own; evidence that it is due to classical conditioning
Is classical conditioning the replacement of the US by the CS?
No, classical conditioning is a learning mechanism where the CS (and the CR) prepare the animal for the onset of the US and UR
Edwards & Acker 1972
Found that WWII veterans reacted to sounds of battle even 15 years after the war
Compensatory reaction hypothesis
Sometimes, the UR and the CR can be opposites: insulin injections (US) deplete blood sugar levels (UR), and after a number of such injections, the body reacts to the CS in an opposite way to how it reacts to the US (blood sugar levels go up as the body ‘prepares’ itself for the injection); could lead to drug overdose if the CS is not present when drug is taken/ administered
Siegel 1989
Tested the tolerance of rats for ‘overdoses’ of heroin in novel or usual environments –> rats more likely to overdose if they were given drug in new as opposed to usual environment where they had drugs (evidence for compensatory reaction hypothesis)
Acquisition
Process by which a conditioned stimulus comes to produce a conditioned response
Trace/ forward conditioning
CS comes before US, but there is a gap between them; not as effective as delayed forward conditioning
Simultaneous conditioning
CS and US start and end together; often fails to produce a CR
Backwards conditioning
CS begins after US; least effective form of classical conditioning
Delayed forward conditioning
Conditioned stimulus comes just before/ overlaps with the unconditioned stimulus; most effective form of classical conditioning
Contingency
How good of a predictor your conditioned stimulus is for the unconditioned stimulus
Contiguity
Learning occurs due to temporal proximity of CS and US
Extinction
If you present the conditioned stimulus without ever presenting the unconditioned stimulus, then the CR would gradually decrease; rate of decrease depends on factors such as initial response strength
Spontaneous recovery
A CS-CR relation is extinguished, however, after a period with no CS presentations, the CS may elicit the CR again
Flooding therapy
Fear elicited by a CS (certain phobias) is eliminated by the process of extinction
Stimulus generalisation
A conditioned response formed to one conditioned stimulus will occur to other, similar stimuli
Generalisation gradients
Stimuli closer to the CS produce greater CRs
Stimulus discrimination
Occurs when an organism does not respond to stimuli that are similar to the stimulus used in training
Discrimination training
Organism is reinforced for responses to one stimulus and not the other –> if organism learns to discriminate, it will respond more to the reinforced stimulus
Systematic desensitisation
Takes a similar stimulus that produces a lesser reaction and repeatedly exposes the subject to it → generalising extinction rather than the acquisition of a response
Blocking
Conditioning does not occur if a good predictor of the US already exists
Higher-order conditioning
Once a stimulus has become an effective CS for a certain CR, then that stimulus can be used to condition other stimuli; effect begins to diminish after a number of trials
Sensory preconditioning
Learning occurs in the absence of UR; classical conditioning reveals the association already learned between two events
Taste-aversion learning
When an individual avoids a certain food or drink due to prior illness/ bad experience with it (substance makes them nauseous)
Thorndike 1874-1935: Law of Effect
Law of Effect: Positive consequences increased the likelihood or probability of a response; behaviours are ‘stamped out’ if followed by negative consequences
Instrumental conditioning
Concerns the probability/ likelihood of a response changing as a result of its consequences; the subject emits the response in order to produce a reward
Skinner 1904-1990
Skinner’s version of the Law of Effect:
When a response is followed by a reinforcer, the strength of the response increases, and when a response is followed by a punisher, the strength of the response decreases
Positive reinforcement
Adding a stimulus or event contingent upon a response increases that behaviour
Negative reinforcement
Removing a stimulus or event contingent upon a response increases that behaviour
Positive punishment
Adding a stimulus or event contingent upon a response decreases that behaviour
Negative punishment
Removing a stimulus or event contingent upon a response decreases that behaviour
Cumulative record
Graph of responses from a conditioning experiment
Continuous reinforcement
Every instance of a response is reinforced; useful for shaping behaviour
Partial or intermittent reinforcement
A designated response is reinforced only some of the time (useful for maintaining behaviours); less likely to undergo extinction because individuals are used to the unreliable nature of the reinforcement
Ratio schedule
Depends on the number of responses, fixed and variable
Interval schedule
Response is still important, but what determines whether or not a response will be reinforced depends on the passage of time
Fixed-ratio schedule
Reinforcer is given after a fixed number of non-reinforced responses
Variable-ratio schedule
Reinforcer is given after a variable number of non-reinforced responses; the number of non-reinforced responses varies around a predetermined average
Fixed-interval schedule
Reinforcer is given for the first response after a fixed period of time has elapsed
Variable-interval schedule
Reinforcer is given for the first response after a variable time interval has elapsed; interval lengths vary around a predetermined average
Gaetani et al. 1986: Engineering compensation systems (effects of commissioned vs. wage payment)
Found that in this particular case, ratio reinforcement (commission-based pay) was more effective than interval reinforcement (hourly pay)
Partial-reinforcement extinction effect
Partial reinforcement schedules provide greater resistance to extinction
Side-effects of extinction:
- Increase in response rate before it goes down
- Increase in response topography (participant takes a different approach in the hopes of being reinforced); example: extinction-induced aggression
What led to the development of Premack’s Principle?
Didn’t take into account how the significance of stimuli can change depending on context (Premack took issue with the idea of trans-situational reinforcers - reinforcers that have the same impact regardless of the situation)
Premack’s Principle
Behaviours are either high probability or low probability; behaviour is reinforced when it is followed by higher probability behaviours
Mitchell & Stoffelmayr 1973
Demonstrated Premack’s Principle with schizophrenic individuals
Honig & Slivka 1964
Found that the effects of punishment are easily generalised
Reynolds 1969
Found that one stimulus dimension had overshadowed learning of the other stimulus dimension
Herrnstein & DeVilliers 1980
Found that non-human animals could form categories or concepts from complex stimuli
The Kelloggs 1933
Raised the chimpanzee Gua alongside their son; Gua learned to understand some commands, but never any English words
The Hayes 1951
Raised Vicki; she learned to make three ‘recognisable’ words (papa, mama, and cup)
1960s attempts at teaching sign language to chimpanzees
Communication possible, but little evidence of syntax
Communication between humans and non-human primates (symbols)
Rumbaugh developed ‘Yerkish’ language for non-human primates; pygmy chimpanzee/ bonobo was tested and results were compared to those of a two-year-old human - had the ability to understand sentences compared to human children’s performance (skill of a two or 2.5-year-old child)
Breland & Breland 1961
Demonstrated biological constraints on instrumental conditioning (pig and raccoon)
Tolman & Honzik 1930
Found that rats actively process information rather than operating on a stimulus-response relationship; latent learning
Observational learning
Occurs when an organism’s response is influenced by the observation of others’ behaviour (models)
Palameta & Lefebvre 1985
Found that group exposed to observational learning was quicker to display the same behaviour (in this case, eating the seed)
Cook & Mineka 1987
Fear of snakes learnt by observation, but also biological constraints - no fear of flowers learned
Bandura et al. (1963, 1965)
Aggression learned through modelling (observational learning)
Bandura et al. 1967
Showing boy playing fearlessly with dog helped to reduce fear (observational learning)
Bandura (four key processes for observational learning):
- Attention: extent to which we focus on others’ behaviour
- Retention: retaining a representation of others’ behaviour
- Production: ability to actually perform actions we observe
- Motivation: need to perform actions we witness (usefulness)
Poche et al. 1988
Observational learning was effective in teaching abduction-prevention skills to children, but not as effective as observational learning paired with rehearsal