Learning Theory Flashcards

You may prefer our related Brainscape-certified flashcards:
1
Q

what is one key assumption of the learning approach?

A

THAT BEHAVIOUR IS MEASURABLE BY USING STIMULUS AND RESPONSE.

THERE IS A FOCUS ON SCIENTITF METHODS.

At the end of the 19th Century people were starting to think about psychological questions. Wundt, an early psychologist, started to use INTROSPECTIVE METHODS which involved participants thinking about their thinking, or how they process information and then explaining it to the researcher.

However, early behavioural theorists believed that we could set up a STIMULUS and OBSERVE what happens as a result of it, this means we can MEASURE RESPONSES quite easily. The assumption therefore leading to the claim that BEHAVIOUR can be studied SCIENTIFICALLY.

About 1900 Thorndike (Behaviourist) based his theory on the idea of the stimulus-response approach – that the stimulus was given to the animal and the response was monitored, which was one of the earliest scientific ways of measuring behaviour that was recorded.

Behaviour is difficult to study and so it is more difficult to draw scientific conclusions in psychology than in chemistry for example, so learning theory psychologists try to isolate a single behaviour and find out what leads to that behaviour.

For example: a single piece of behaviour (a rat pressing a lever for a food pellet) is OBSERVED to see how quickly the rat will learn to press the lever for a reward if a red light flashes a signal the rat will get food a green light signals there is no food.

THAT BEHAVIOUR IS MEASURABLE BY USING STIMULUS AND RESPONSE.

THERE IS A FOCUS ON SCIENTITF METHODS.

At the end of the 19th Century people were starting to think about psychological questions. Wundt, an early psychologist, started to use INTROSPECTIVE METHODS which involved participants thinking about their thinking, or how they process information and then explaining it to the researcher.

However, early behavioural theorists believed that we could set up a STIMULUS and OBSERVE what happens as a result of it, this means we can MEASURE RESPONSES quite easily. The assumption therefore leading to the claim that BEHAVIOUR can be studied SCIENTIFICALLY.

About 1900 Thorndike (Behaviourist) based his theory on the idea of the stimulus-response approach – that the stimulus was given to the animal and the response was monitored, which was one of the earliest scientific ways of measuring behaviour that was recorded.

Behaviour is difficult to study and so it is more difficult to draw scientific conclusions in psychology than in chemistry for example, so learning theory psychologists try to isolate a single behaviour and find out what leads to that behaviour.

For example: a single piece of behaviour (a rat pressing a lever for a food pellet) is OBSERVED to see how quickly the rat will learn to press the lever for a reward if a red light flashes a signal the rat will get food a green light signals there is no food.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

what is the second key assumption of the learning approach?

A

THAT OUR ENVIRONMENT SHAPES OUR BEHAVIOUR.

THERE IS A FOCUS ON THE ENVIRONMENT.

Our behaviour is argued in psychology to come from either INHERITED (nature) or ENVIRONMENTAL (nurture) influences.

Watson claimed that if all our behaviour comes from conditioning and reinforcement between certain stimuli in the environment, then it is the environment that shapes us, and not inherited characteristics. Our environment is everything that we experience, e.g. school, culture, family etc.

The learning approach, by studying what causes an action (the stimulus) and the action itself (the response) assumes that behaviour arises from experiences around us, not from forces within us.

Actions do not come from brain activity, or from our inherited genetic abilities, but from experiences that we have. For example, a baby will babble “m..m..m..” and the mother will respond “mummy” and give the child attention (reinforcement) so the child will repeat the sound and eventually learns to talk. Punishment can also shape behaviour, people are less likely to repeat a behaviour if they are punished for it.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

general principle of classical conditioning:

A

Classical conditioning focuses on STIMULUS RESPONSE, the conditioning of REFLEXES. Human reflexes include a fear response, eye blinking, knee-jerking, fear and breathing. Fear can be conditioned, as fear is a reflexive response, this can explain phobias.

Most behaviour goes beyond just a reflex so classical conditioning is quite limited.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

general principle of operant conditioning:

A

Proposes that behaviour is for most part VOLUNTARY rather than reflexive and so operant conditioning looks at how reinforcement and punishment shapes behaviour. For example, school children work hard for a reward of gold starts, so their behaviour (working hard) is repeated for the REINFORCEMENT (a gold star). Some children might not work hard and be PUNISHED by being given detention.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

general principles of social learning theory:

A

Social Learning Theory was first proposed in 1960. It involves people learning through OBSERVATION of others. Role models (someone you look up to or admire) are more favourable to observe and copy. SLT therefore proposes that not all behaviour is reinforced (operant conditioning) or based on stimulus-response reflexes (classical conditioning), and that some behaviour is based on what we see around us.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

overview of classical conditioning:

A

Classical conditioning is a theory of learning that examines how a response is associated with a stimulus to cause conditioning.

A stimulus is something that produces a response, which in classical conditioning is either a reflex or an automatic behaviour. These responses to stimuli are involuntary responses i.e. showing a startle response like fear to a sudden noise.

Pavlov originally trained as a medical doctor and conducted research on the nervous system and digestion which won him the Nobel Prize in 1904. He established the world’s first clinic and operating theatre to be used exclusively for animals which was maintained to the highest standard because he realised that he would not get reliable results if the animals were suffering or in distress from other reasons than their illness or what was being done to them.

Pavlov deduced that dogs responses behaved in ways that showed anticipation of the situation and based his theory of classical conditioning on this principle.

Classical conditioning argues that there is an association between an UNCONDITIONED STIMULUS (UCS), which is a naturally occurring stimulus, and an existing UNCONDITIONED RESPONSE (UCR), which is a naturally occurring response.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

describe extinction is classical conditioning:

A

Refers to when the association (CS and CR) is no longer there, it is extinguished. So when the association is extinguished the bell (CS) will no longer cause salivation (CR). This can be achieved by presenting the UCS (food) without the CS (bell). The association between the two stops, and therefore the bell returns to being a neutral stimulus. So, the UCS of food presented without CS of bell results in CR of salivation. The CS of bell will then begin to result in no response, therefore the bell returns to original state of NS.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

describe spontaneous recovery in classical conditioning:

A

This is when, after extinction, you might find that a previously paired conditioned response to a conditioned stimulus suddenly and without reconditioning reappears for no reason. For example above, despite extinction, you ring a bell and the response of salivation occurs again.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

describe stimulus generalisation in classical conditioning:

A

Is another term of classical conditioning is stimulus generalisation, which refers to the extending of the original association between a CS and CR to include one or more similar stimulus. For example, if a dog becomes conditioned to salivate to a particular bell sound (CS), then it might generalise and salivate to some other bell sounds (CR).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

describe discrimination in classical conditioning:

A

This is the opposite of stimulus generalisation. It is when the conditioning is associated with only one specific conditioned stimulus, and only that specific CS can result in the conditioned response. For example, having a fear of pink buttons but not any other type of button shows a discrimination between the types of button.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

describe pavlol salivating dogs experiment:

A

Prior to the experiment, the dogs would salivate (UCR) in response to the meat powder food but not to the sound. During the conditioning phase the meat powder was presented at the same time as a metronome, with repeated pairings of meat powder and metronome CONDITIONING resulted.

As a result of the pairings, the NS (the metronome) had become a CONDITIONED STIMULUS (CS) capable of producing the behaviour of salivating which is a CONDITIONED RESPONSE (CR).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

2 strengths for classical conditioning:

A

Olson and Fazio (2001) used a laboratory study and found that classical conditioning contributes to the development of some gender-related attitudes therefore providing support for stimulus response behaviour.

Classical Conditioning can explain the acquisition of some aspects of behaviour, where a particular response is associated with a particular stimulus e.g. Gulf War Syndrome can also be explained as a learned reaction to the war environment.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

2 weaknesses of classical conditioning:

A

Pavlov studied non-human animals and then applied their findings to humans which are difficult to generalise because animal cognitive and biological systems are not the same as humans.

Classical conditioning proposed that behaviour is the result of stimulus-response association, this is not the only explanation of human behaviour, social learning theory suggests that it is due to imitation and modelling and not stimulus-response.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

overview of operant conditioning:

A

Operant conditioning relates to learning through consequences. The idea is that if you are given a consequence for a specific behaviour you will either repeat it, or stop it.

Operant conditioning differs from classical conditioning because the consequences come after the behaviour whereas in classical conditioning the behaviour (reflex) comes before or at the same time as the stimulus.

THORNDIKE (1911)

Check your notes about Thorndike on page 6 of this booklet.

Thorndike (1911) researched reinforcement using a kitten in a puzzle box. Initially when the kitten was placed in the box it behaved randomly, however once it accidentally hit a lever that opened the door it received food.

After several trials the kitten escaped from the puzzle box faster, pressing the lever to open the door immediately – demonstrating that it had learned by trial and error.

This is Thorndike’s (1911) Law of Effect; a behaviour (pressing lever) followed by a pleasant consequence (food) tends to be repeated, while a behaviour (pressing button) followed by an unpleasant consequence (loud noise) tends not to be repeated.

Such a procedure differs from Pavlov’s classical conditioning because the cat only received the food as a consequence of performing the desired behaviour.

The increased success of the kitten opening the box for the food is shown in the graph below. The increased success of opening the box for food is LATENCY

This reduction in LATENCY time from about five minutes to as little as five seconds over the trials indicated that the cats were learning the puzzle.

SKINNER (1930)

In order to test Thorndike’s ideas further and to maximize the OBJECTIVITY, ACCURACY and EASE of recording behaviour in experiments, Skinner developed a specially made chamber in which an animal, e.g. a rat or pigeon, could learn a specific response.

The apparatus, a Skinner Box, could present stimuli (known as ANTECEDENT) which allowed responses (BEHAVIOUR) to be measured and recorded. CONSEQUENCES followed the performance of a particular desired behaviour.

The antecedents (a) included lights and noises, the behaviour (B) where the animal presses or pecks at a disc and the consequences (C) were food or electric shocks.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

describe reinforcement and punishment concepts:

A

Reinforcement refers to a consequence of a behaviour, used to encourage repetition.

Punishment refers to a consequence of a behaviour, used to prevent repetition

Positive means to receive (i.e. to add +) a consequence.

Negative means to remove (i.e. to take away -) a consequence.

POSITIVE AND NEGATIVE DO NOT, NEVER HAVE, NEVER WILL AND NEVER SHALL BE ASSOCIATED WITH THE DESIREABILITY OF THE BEHAVIOUR SHOWN! IT’S ABOUT ADDING AND TAKING AWAY (+ AND -) NOT WHETHER YOU ARE GOOD OR BAD!

POSITIVE REINFORCEMENT is when something desired is received in response to a particular behaviour. So, if a child tidies their room, they receive (+ positive) extra pocket money. They have been positively reinforced to tidy their rooms because they desire the consequence of receiving pocket money.

POSITIVE REINFORCEMENT ENCOURAGES REPETITION OF BEHAVIOUR.

NEGATIVE REINFORCEMENT is when something undesired is taken away in response to a particular behaviour. So, if a person does not like the loud music in a particular restaurant (undesired), they will go to a quieter restaurant because they want to remove (- negative) the loud music. They have been negatively reinforced to go to the other restaurant because they desire the consequence of removing the music.

NEGATIVE REINFORCEMENT ENCOURAGES REPETITION OF BEHAVIOUR.

POSITIVE PUNISHMENT is when something unpleasant is received (+ positive) in response to a particular behaviour. So, a child who did not complete homework (behaviour) might be given (+ positive) detention. They have been positively punished for not doing homework because the consequence of receiving detention is undesired.

POSITIVE PUNISHMENT DISCOURAGES REPETITION OF BEHAVIOUR.

NEGATIVE PUNISHMENT is when something pleasant is removed (- negative) in response to a particular behaviour. So, a child who did not complete homework (behaviour) might have their phone taken off them, so removing (- negative) something they enjoy. They have been negatively punished for not doing homework because the consequence of removing their phone is undesired.

NEGATIVE PUNISHMENT DISCOURAGES REPETITION OF BEHAVIOUR.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

describe skinners 1948 superstition in the pigeon:

A

Skinner et al (1968) highlighted the importance of a CONDITIONED RESPONSE – that the reinforcement should be dependent on the behaviour. This suggests that when a positive consequence follows behaviour we tend to repeat that behaviour, this repetition of behaviour happens even if it was not the actually cause of the reward.

Positive consequences that occur regardless of our behaviour are called uncontrollable reinforcers. They give rise to superstition. The behaviours that arise as a consequence of such reinforcement are called SUPERSITIOUS BEHAVIOURS.

We could expect that superstition, the belief in a relationship which does not in fact exist, to be uniquely human as we assume that animals do not have beliefs. However, Skinner designed this experiment to demonstrate that superstitious behaviours could be acquired by animals. He proposed that superstitious behaviours simply arise because they are accidentally reinforced.

TASK: summarise the information about the Pigeon study.

MATUTE (1996) tested the role of uncontrollable reinforcers in human behaviour and tested whether a superstitious behaviour can arise because we fail to test the possibility that a positive outcome can occur irrespective of our behaviour.

Participants worked on computers, some of which periodically emitted beeps. If the participants had the opportunity to discover that doing nothing led to the beep stopping on its own, they learned to wait. However, when they were instructed to try and stop the beep, the participants acquired superstitious responses. In an attempt to stop the noise they tried pressing particular buttons, and learned (wrongly) that it was this button that made the noise stop.

17
Q

describe continuous reinforcement:

A

Continuous Reinforcement (CR) offers a reward every time the behaviour is completed. Other schedules of reinforcement are only PARTIAL, which will ultimately affect both response rate and resistance to extinction.

18
Q

describe fixed interval reinforcement:

A

A fixed interval reinforcement schedule is when behaviour is rewarded after a set amount of time. For example, June undergoes major surgery in a hospital. During recovery, she is expected to experience pain and will require prescription medications for pain relief. June is given an IV drip with a patient-controlled painkiller. Her doctor sets a limit: one dose per hour. June pushes a button when pain becomes difficult, and she receives a dose of medication. Since the reward (pain relief) only occurs on a fixed interval, there is no point in exhibiting the behaviour when it will not be rewarded.

19
Q

describe variable interval reinforcement:

A

With a variable interval reinforcement schedule, the person or animal gets the reinforcement based on varying amounts of time, which are unpredictable. Say that Manuel is the manager at a fast-food restaurant. Every once in a while someone from the quality control division comes to Manuel’s restaurant. If the restaurant is clean and the service is fast, everyone on that shift earns a $20 bonus. Manuel never knows when the quality control person will show up, so he always tries to keep the restaurant clean and ensures that his employees provide prompt and courteous service. His productivity regarding prompt service and keeping a clean restaurant are steady because he wants his crew to earn the bonus.

20
Q

describe variable ratio reinforcement:

A

In a variable ratio reinforcement schedule, the number of responses needed for a reward varies. This is the most powerful partial reinforcement schedule. An example of the variable ratio reinforcement schedule is gambling. Imagine that Sarah—generally a smart, thrifty woman—visits Las Vegas for the first time. She is not a gambler, but out of curiosity she puts a quarter into the slot machine, and then another, and another. Nothing happens. Two dollars in quarters later, her curiosity is fading, and she is just about to quit. But then, the machine lights up, bells go off, and Sarah gets 50 quarters back. That’s more like it! Sarah gets back to inserting quarters with renewed interest, and a few minutes later she has used up all her gains and is $10 in the hole. Now might be a sensible time to quit. And yet, she keeps putting money into the slot machine because she never knows when the next reinforcement is coming. She keeps thinking that with the next quarter she could win $50, or $100, or even more. Because the reinforcement schedule in most types of gambling has a variable ratio schedule, people keep trying and hoping that the next time they will win big. This is one of the reasons that gambling is so addictive—and so resistant to extinction.

20
Q

describe fixed ratio reinforcement:

A

With a fixed ratio reinforcement schedule, there are a set number of responses that must occur before the behaviour is rewarded. Carla sells glasses at an eyeglass store, and she earns a commission every time she sells a pair of glasses. She always tries to sell people more pairs of glasses, including prescription sunglasses or a backup pair, so she can increase her commission. She does not care if the person really needs the prescription sunglasses, Carla just wants her bonus. The quality of what Carla sells does not matter because her commission is not based on quality; it’s only based on the number of pairs sold. This distinction in the quality of performance can help determine which reinforcement method is most appropriate for a particular situation. Fixed ratios are better suited to optimize the quantity of output, whereas a fixed interval, in which the reward is not quantity based, can lead to a higher quality of output.

21
Q

describe the application to gambling:

A

Skinner (1953) stated, “If the gambling establishment cannot persuade a patron to turn over money with no return, it may achieve the same effect by returning part of the patron’s money on a variable-ratio schedule”.

Skinner uses gambling as an example of the power and effectiveness of conditioning behavior based on a variable ratio reinforcement schedule. In fact, Skinner was so confident in his knowledge of gambling addiction that he even claimed he could turn a pigeon into a pathological gambler (“Skinner’s Utopia,” 1971). Beyond the power of variable ratio reinforcement, gambling seems to work on the brain in the same way as some addictive drugs. The Illinois Institute for Addiction Recovery reports evidence suggesting that pathological gambling is an addiction similar to a chemical addiction. Specifically, gambling may activate the reward centers of the brain, much like cocaine does. Research has shown that some pathological gamblers have lower levels of the neurotransmitter (brain chemical) known as norepinephrine than do normal gamblers (Roy, et al., 1988). According to a study conducted by Alec Roy and colleagues, norepinephrine is secreted when a person feels stress, arousal, or thrill; pathological gamblers use gambling to increase their levels of this neurotransmitter. Another researcher, neuroscientist Hans Breiter, has done extensive research on gambling and its effects on the brain. Breiter reports that “Monetary reward in a gambling-like experiment produces brain activation very similar to that observed in a cocaine addict receiving an infusion of cocaine”. Deficiencies in serotonin (another neurotransmitter) might also contribute to compulsive behavior, including a gambling addiction.

It may be that pathological gamblers’ brains are different than those of other people, and perhaps this difference may somehow have led to their gambling addiction, as these studies seem to suggest. However, it is very difficult to ascertain the cause because it is impossible to conduct a true experiment (it would be unethical to try to turn randomly assigned participants into problem gamblers). Therefore, it may be that causation actually moves in the opposite direction—perhaps the act of gambling somehow changes neurotransmitter levels in some gamblers’ brains. It also is possible that some overlooked factor, or confounding variable, played a role in both the gambling addiction and the differences in brain chemistry.

However, some research suggests that pathological gamblers use gambling to compensate for abnormally low levels of the hormone norepinephrine, which is associated with stress and is secreted in moments of arousal and thrill.