Chapter 5 Flashcards

1
Q

Operant conditioning is a form of ____ learning.

A

Associative.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

In classical conditioning, do we control the response?

A

No.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Classical VS operant conditioning?

A

In classical the outcome occurs regardless, whilst in operant the outcome is dependent on your response.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Operant conditioning is based on avoiding or obtaining a specific ____.

A

Outcome.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Operant conditioning requires an ____ operate in it’s environment to determine an outcome.

A

Organism.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Thorndike was the first to study behavioral outputs due to operant conditioning. What was his study? What was the conclusion? What is this idea known as?

A

Puzzle boxes. Organisms are more likely to repeat actions that produce satisfying consequences, and less likely to repeat actions that do not. This idea is known as law of effect.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

The more time an animal spent in Thorndike’s box, the ____ they learned to escape.

A

Quicker.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Law of effect?

A

Probability that a particular behavioral response increases or decreases depending on the consequences that have followed that response in the past.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

____ -> ____ -> ____

A

Stimulus, response, outcome.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Thorndike’s learning procedures involved ____ trials. What is this?

A

Discrete. Operant conditioning paradigm whee the experimenter defines the beginning and end of each trial.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

BF Skinner wanted to refine Thorndike’s techniques. How did he do this?

A

He created a Skinner box, which was opposite from Thorndike’s discrete trial. The Skinner box is a conditioning chamber where reinforcement/punishment is automatically delivered when an animal makes a response (ex: lever pressing) - in this case, the animal is in charge of what was the start an end.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Free Operant Paradigm?

A

Operant conditioning paradigm where the animal can operate the apparatus “freely”, responding to obtain reinforcement/avoid punishment, whenever it chooses - commonly referred to as operant conditioning

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Can we see extinction in operant conditioning?

A

Yes.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Reinforcement?

A

Providing consequences to increase probability of behavior occurring again in the future.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Punishment?

A

Providing consequence to decrease probability of behavior occurring again the in the future.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Adding a stimulus to free operant experiments can make them more elaborate. Example?

A

S(Light ON)->R(Lever press)->O(Food release)

S(Light OFF)->R(Lever press)->O(NO food release)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

According to Thorndike and Skinner, operant conditioning consists of what 3 components?

A

Stimulus, response, outcome.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

Discriminative stimuli?

A

In operant conditioning, stimuli that signal whether a particular response will lead to a particular outcome.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

Example of Shaping?

A

Little kid learning to write the letter “A”, it looks somewhat similar and they are rewarded it. Then he kept getting closer and closer to a proper A. Shaping: operant conditioning technique in which successive approximations to a desired response are reinforced.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

Chaining? Example?

A

Operant conditioning technique where organisms are gradually trained to execute complicated sequences of discrete responses. I.e. Learning a complicated dance in order to receive smarties? How about you get a smartie after every correct dance move.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

Operant conditioning?

A

Process whereby organisms learn to make responses in order to obtain or avoid important consequences

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

Operant conditioning

A

Process whereby organisms learn to make our refrain from making certain responses in order to obtain/avoid a certain outcome.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

Thorndike’s cat in box

A

A cat was placed in a box and when it escaped it was given food so it was more likely to do it again

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
24
Q

Law of effect

A

Given a particular stimulus, a response that leads to a desirable outcome will tend to increase in frequency

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
25
Q

Classical vs operant conditioning

A

Classical: organisms experience an outcome (US) whether or not they have learned the conditioned response (CR)
Operant: the outcome O is wholly dependent on whether the organism performs the response R

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
26
Q

How are operant and classical conditioning similar?

A

Both have a negatively accelerated learning curve (time to do something decreases rapidly and then levels off) and they both show extinction

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
27
Q

Discrete trial paradigms

A

Experimented defined the beginning and end of each trial

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
28
Q

Free operant paradigm

A

The animal can operate freely (i.e running into maze to get food, then running out and that’s when the trial ends.)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
29
Q

Reinforcement does what?

A

Increase probability of behavior

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
30
Q

Skinner box

A

Cage with lever for food to be dispensed into a little plate thing. The animals would explore the cage and they accidentally hit it and dramatically increased their rate of responding

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
31
Q

Punishment does what?

A

Decrease probability of a certain behavior

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
32
Q

Cumulative recorder

A

height of the line at any given time represents the number of responses that have been made in the entire experiment (cumulative) up to that time

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
33
Q

Discriminative stimuli

A

Stimuli that signal whether a particular response will lead to a particular outcome

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
34
Q

Habit slip

A

Stimulus to response association is so strong that the stimulus seems to evoke the learned response automatically i.e. Making a phone call to a familiar number instead of the number you intended to dial

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
35
Q

Protestant ethic effect

A

Rats that have been trained to press a lever to obtain food will often continue to work to obtain food even though they have free food in their cage

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
36
Q

What is a response defined by?

A

The outcome it produces

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
37
Q

Shaping: successive approximations to the desired response are reinforced

A

When a rat in a Skinner box happens to wander near the food tray, experimenter drops in a piece of food. The rat eats the food and starts to learn an association b/w the tray and food and will spend all its time near the food tray. The experimenter then changes the rules: now rat must also be near the lever before food is dropped. Once the rat has learned this, rules change again: food is dropped only if the animal is actually touching the lever, then rule changes again: only if animal is pressing down the lever.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
38
Q

A reinforcer is a ____.

A

Consequence

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
39
Q

Primary reinforcer

A

Organisms have innate drive to obtain these things i.e. Food water sex sleep good temperature

40
Q

Clark Hull’s drive reduction theory

A

Drive reduction theory says that humans are motivated to reduce the state of tension caused when certain biological needs are not satisfied.

41
Q

what are some complications with primary reinforcements?

A

An animal will work hard for water, but once they’ve drunk enough, more water isn’t reinforcing. Also, primary reinforcers are not created equally (animal will work harder for food they like)

42
Q

Secondary reinforcers

A

Reinforcers that initially have no intrinsic value but that have been paired with primary reinforcers i.e. Money

43
Q

Token economies

A

Desired behavior is reinforced with a token which can be exchanged for privilege

44
Q

By being paired with a primary reinforcer, secondary reinforcers become reinforcers themselves.

A

Organisms will work to obtain.

45
Q

Why do some researchers say that animals aren’t fooled by secondary reinforcers?

A

They think the animals use the secondary reinforcers for providing informational feedback that behavior is on the right track for obtaining a primary reinforcer

46
Q

A switch in outcome may produce a change in ____.

A

Responding.

47
Q

Negative contrast?

A

organisms given a less preferred reinforcer in place of an expected and preferred reinforcer will respond less strongly for the less preferred reinforcer than if they had been given that less preferred reinforcer all along

48
Q

Negative contrast example

A

while rats can be trained to make lever press responses to obtain either food pellets or water sweetened with sucrose, they tend to prefer the latter. If the sweetened water is used as reinforcer during the first half of each training session and food pellets as the reinforcer during the second half, rats typically make many more responses during the first half of the session

49
Q

Thorndike and Skinner concluded that punishment was _______ as reinforcement at controlling behavior

A

Not as effective

50
Q

How can discriminative stimuli for punishment encourage cheating?

A

A speeding driver will see a cop (the discriminative stimulus) and slow down to avoid getting a ticket. But speeding in the absence of a police car will probably not be punished. In this case, punishment doesn’t train driver not to speed. It only teaches him to suppress speeding in the presence of police cars. In effect, driver has learned to cheat

51
Q

How can concurrent reinforcement undermine punishment?

A

The effects of punishment can be counteracted if the reinforcement occurs along with the punishment. Like suppose a rat first learns to press a lever for food but later learns that lever presses are punished by shock. Unless the rat has another way to obtain food, it’s likely to keep pressing the lever to obtain food reinforcement in spite of punishing effects of shock.

52
Q

How can punishment not end how you want?

A

Punishment decreases the probability that R will happen in the future… but what will happen instead? The organism will explore possible responses

53
Q

Does initial intensity of punishment matter?

A

Yes it should be strong. I.e. Rats become numb to mild shock

54
Q

Reinforcement schedules

A

Rules determining when outcomes are delivered in an experiment

55
Q

____ outcomes produce fastest learning.

A

Immediate.

56
Q

Response -> outcome interval being long results in what?

A

Decrease in association

57
Q

The time lag between R and O is an important factor in ____.

A

Self control e.g. its easy to convince a student to study if a test is coming up tomorrow; it is harder if the test is in 5 weeks. the delay between R (studying) and O (good grades) makes reinforcement less effective in eliciting the response

58
Q

Pre commitment

A

Making a choice that is difficult to change later. For example a dieter will be less likely to cheat if he gets rid of all the sweets in their house

59
Q

Positive reinforcement versus positive punishment

A

Positive reinforcement
S (potty present) –> R (empty bladder) –> O (praise)
performance of the response causes the reinforcer to be “added”
Positive Punishment
S (potty absent) –> R (empty bladder) –> O (Disapproval)
The response must be with held, if it is not withheld (toddler peeing himself) a punishment is “added” to the environment

60
Q

Negative reinforcement versus negative punishment

A

Negative Reinforcement
S (shock) –> R (take Aspiring) –> O (no more headache)
behaviour is encouraged because it causes something negative to be subtracted from the environment
Negative Punishment
S (recess) –> R (aggressive behaviour) –> O (loss of playtime)
Behaviour is discouraged by subtracting something reinforcing (playtime) from the environment

61
Q

Continuous reinforcement schedule

A

Each response is always followed by the outcome

62
Q

Partial reinforcement schedules

A

Patterns where response is followed by outcome less than 100% of the time

63
Q

Fixed ratio schedule

A

A fixed number of responses must be made before outcome is delivered

64
Q

Fixed interval schedule

A

Reinforcers first response after a fixed amount of time. Once the time interval has elapsed, the reinforcement remains available until the response occurs and the reinforcement is obtained. To constantly respond before the times interval elapses is a waste of time and effort

65
Q

Variable ratio schedule

A

Provides reinforcement after random number of responses. This reduced the post reinforcement pause because there’s a study higher rate of responding since the very next response my result in a reinforcement (i.e. Lottery)

66
Q

Variable interval schedule

A

Reinforce his first response after an interval that averages in particular length in time. Response rate under variable interval is steadier then fixed interval because animals check periodically to see whether reinforcement is available

67
Q

Concurrent reinforcement schedule

A

Organism can make any of several possible responses, each leading to a different outcome. This allows researchers to examine how organisms choose to divide your time and efforts among different options

68
Q

Matching law of choice behavior

A

Behavior is correlated with it’s environment. Given two responses that are reinforced on variable interval schedules, an organisms relative rate of making each response will match the relative rate of reinforcement for that response

69
Q

Behavioral economics

A

Study of how organisms allocate their time and resources among possible options

70
Q

Economic theory predicts that each consumer will allocate resources in a way that maximizes there____or relative satisfaction

A

Subjective value

71
Q

The particular allocation of resources to provide maximum subjective value to an individual is called a____

A

Bliss point. I.e. Jamie gets $100 each week, he can either buy 10 albums or eat out five times, is bliss point is eating out two times a week but buying six albums

72
Q

The premack principle

A

Opportunity to perform a highly frequent behavior can reinforce a less frequent behavior. For example, peanut gave a group of rats free access to drinking water and the running wheel. On average, right spent more time running than drinking. He restricted their access to the wheel: they were allowed to run after the drunk certain amount of water. They learned the response to outcome association and started drinking more water (R) in order to access the wheel (O). Total amount of running decreased and total amount of drinking increased. Activity of running was acting as a reinforcer, and it was increasing the probability of an infrequent behavior

73
Q

Response deprivation hypothesis

A

Suggests that the critical variable is not which responses more frequent, but merely which response have been restricted. By restricting the ability to make any response, you can make the opportunity to perform that response more reinforcing. For example, if you have been studying for six hours straight, The idea of taking a break to clean your room (which is something you hate) may begin to look attractive

74
Q

Cortical areas process____information and then send it as primary inputs to the____, which sends messages to motor neurons in the muscles.

A

Sensory, motor cortex

75
Q

Dorsal striatum

A

Part of the basal ganglia, is further divided into caudate nucleus and putamen. Receive stimulus information from sensory cortical areas and projects to the motor cortex.

76
Q

Dorsal striatum is critical for what?

A

It is critical in operant conditioning for learning stimulus to response associations based on feedback. For example, rats with lesions of the DS can learn simple response to outcome associations but only in the absence of discriminative stimuli

77
Q

Orbitoftontal cortex

A

One of the brain areas that appear to be important in prediction of behavior. This area is underneath front of the brain that contributes to goal directed behavior by representing predicted outcomes. It receives input from all sensory modalities. Monkeys with OFC lesions can learn to respond to stimuli that have a reward outcome over a stimuli that does not have a reward outcome. But, if conditions are reversed and the second stimulus is now the one being rewarded, lesioned monkeys are slower to adapt

78
Q

Neurons in the orbifrontal cortex fire differently depending on what?

A

Whether a reward or punishment is coming. Medial areas of OFC we process information about reinforcers and lateral aerials about punishers.

79
Q

Ventral tegmental area (VTA) is known as the…?

A

Pleasure centre

80
Q

In humans, pimizide suppresses cravings, whilst ____ increases cravings.

A

Dopamine

81
Q

Dopamine also strengthens learning of stimulus to response associations during operant conditioning by generally promoting____.

A

Synaptic plasticity

82
Q

Opiates signal ____ in the brain.

A

Liking

83
Q

Endogenous opioids?

A

Naturally occurring peptides with similar effects to opiates. Researchers believe drugs like heroin or so pleasurable because the activate the same brain receptors as endogenous opioids. Endogenous opioids are released in response to primary reinforcers

84
Q

What is wanting signaled by?

A

Dopamine

85
Q

What is liking signaled by?

A

Endogenous opioids

86
Q

How can you differ addiction from habit?

A

The degree of it

87
Q

What do pathological addicts experience?

A

Extreme withdrawal symptoms, to the point where the neglect other parts of their life because nothing else is as important as the addictive substance.

88
Q

Addiction involves not only seeking the high, positive reinforcement, but also avoiding the adverse effects of withdrawal,____, both of which reinforce drunk taking.

A

Negative reinforcement

89
Q

Many addictive drugs are___, which target opiate receptors, which increase dopamine levels

A

Opiates

90
Q

What does cocaine block,?

A

Dopamine reuptake, so it stays in the synapse longer

91
Q

Many addicts report that they no longer experience a high to cocaine and amphetamines, why is that?

A

Their wanting system has disconnected from their liking system.

92
Q

Behavioral addiction

A

When addictions to behaviors, instead of drugs, produce reinforcements or highs as well as cravings and withdrawal symptoms when the behavior is prevented. The most common example of this is gambling. Skaters adjusted gambling is so addictive due to reinforcement on variable ratio schedule.

93
Q

Where does behavioral addiction reflect dysfunction in the brain?

A

The same place that is affected by drug addictions. For example gambling activates the brain similarly to how cocaine does.

94
Q

In the US, most treatment plans include what type of therapy?

A

Cognitive therapy

95
Q

Addiction is a strong stimulus to response to Alcom Association, with environmental stimuli triggering the addictive behavior, resulting in the reinforcing outcome of the high. The’s, treatment should reduce the ____ strength.

A

True

96
Q

Distancing

A

Avoiding the stimuli the trigger the unwanted response

97
Q

Delayed reinforcement

A

Imposing a fix delay before giving into addiction. Delete between response and I’ll come weekends learning of response to outcome response