Topic 4: Operant conditioning: reinforcement Flashcards

1
Q

Operant (instrumental) conditioning

A

Learning that is controlled by the consequences of the organisms behavior

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Reinforcement

A

-Process in which a behavior is strengthened by the immediate consequence that reliably follows its occurrence
-Strengthened=more likely to occur in future
-Thorndike law of effect
-Skinner + operant boxes

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Thorndike’s law of effect

A

-“If a response, in the presence of a stimulus, is followed by a satisfying state of affairs, the bond between stimulus and response will be strengthened”
-Satisfaction=Stamping in
-Discomfort=Stamping out

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Positive reinforcement

A

-Adding something good/desirable that makes the behavior occur in the future
-Eg) Yelling at a kid who enjoys the attention at being yelled at will yell again in the future
-Adding in yelling (positive in this scenario) leads to more yelling in future (behavior reinforced)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Negative reinforcement

A

-Taking away something bad/undesirable that encourages behavior to occur more often
EG) Turning off a loud buzzing noise when child finishes cleaning room
-Negative reinforcement does NOT equal punishment

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Positive punishment

A

-Adding something bad/undesirable to make behavior happen less often
-EG) adding more chores when child misbehaves

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Negative punishment

A

-Taking away something good that leads to behavior occurring less in the future
-EG) kid is having fun but is too loud, they get a timeout away from having fun. When out of timeout they wont be loud again to avoid being sent away from having fun

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Antecedent

A

-The conditions you are in that determines whether a behavior will occur or not
-Could also be called stimulus

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Operant behavior

A

-A behavior that is strengthened through the process of reinforcement
-Acts on environment to produce a consequence

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Operant learning

A

Change in a behavior as a function of the consequence that followed it

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Reinforcement

A

The procedure of providing consequences for a behavior that increases or maintains the probability of that behavior occurring in the future

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Reinforcer

A

Any event or stimulus that follows an operant response and increases or maintains its future probability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Positive reinforcement #2

A

Any event or stimulus that, when presented as a consequence of a behavior, increases or maintains the future probability of that behavior

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Negative reinforcement #2

A

Any event or stimulus that, when removed as a consequence of a behavior, increases or maintains the future probability of that behavior

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Escape behavior

A

-When operant behavior increases by removing an ongoing event/stimulus
-Eg) turning off alarm clock or pressing lever to stop electric shock

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Avoidance behavior

A

-When operant behavior increases by preventing the onset of the event or stimulus
-Eg) pressing a lever to prevent an electric shock

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

Discrete trial procedure

A

-Instrumental response produced once per trial
-Each training trial ends with removal of animal from the apparatus
-Each trial is done as an isolated chunk

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

Free-Operant procedure

A

-Animals remain in apparatus and can make many responses
-No intervention by the experimenter
-Developed by skinner
-Continuous trials

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

Cumulative record

A

-Based on old cumulative recorder device
-Constant paper output, pen jumps with each
-Plot of cumulative responses (y-axis) over time (x-axis)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

Unconditional (primary) reinforcer

A

-A reinforcer that acquired its properties as a function of species evolutionary theory
-Stimuli and events have biological importance
-Usually depends on some amount of deprivation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

conditional reinforcer

A

-Otherwise neutral stimuli or events that have acquired the ability to reinforce due to a contingent
relationship with other, typically unconditional, reinforcers

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

Immediacy

A

-A stimulus is more effective as a reinforcer when it is delivered immediately after the behavior

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

Specific reinforcer used

A

-Certain reinforcers are preferred over others
-Chocolate > Sunflower seeds

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
24
Q

Task characteristics

A

Reinforce a pigeon pecking for food vs. a hawk pecking for food

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
25
Q

Contingency

A

-A stimulus is more effective as a reinforcer when it is delivered contingent on the behavior
-The degree of correlation between a behavior and its consequence

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
26
Q

Contiguity

A

-Nearness of events in time (temporal contiguity) or space (spatial contiguity)
-High contiguity referred to as pairing
-Less contiguity (longer delays) between the operant responses and the reinforcer diminishes the effectiveness of the reinforcer
-Hyperbolic decay function

27
Q

Motivating operations

A

-Establishing operations
-Abolishing operations

28
Q

Establishing operations

A

-Make a stimulus more effective as a reinforcer at a particular time
-Eg) deprivation

29
Q

Abolishing operations

A

-Make a stimulus less potent as a reinforcer at a particular time
-Eg- Satiation

30
Q

Reinforcer magnitude

A

-Generally a more intense stimulus is a more effective reinforcer
-relationship between size and effectiveness is not linear
-More magnitude increases the less benefit you get from increase
-Effectiveness of unconditional reinforcers tend to diminish quickly

31
Q

Schedule of reinforcement

A

-A rule describing the delivery of reinforcement
-Different schedules produce unique schedule effects

32
Q

Schedule effects

A

-Particular pattern and rate of behavior over time
-Over the long-term, effects are very predictable
-Occur in numerous species (humans too)

33
Q

Continuous reinforcement schedule

A

-Behavior is reinforced each time it occurs
-Rate of behavior increases rapidly
-rare in natural environment

34
Q

intermittent reinforcement schedule

A

-Four different types
1) fixed ratio
2) Variable ratio
3) Fixed interval
4) Variable-interval

35
Q

1) fixed ratio schedule

A

-Behavior reinforced after a fixed number of times
-Generates post-reinforcement Pause
-Generates study run rates followed the post-reinforcement pause (PRP)

36
Q

Post-reinforcement pause

A

Pausing typically increases the ratio size and reinforcer magnitude

37
Q

2) Variable ratio schedule

A

-Number of responses needed varies each time
-Ratio-requirement varies around an average
-PRP’s are rare and very short, they are influenced by the lowest and/or average ratio
-Produces higher rates than a comparable fixed-ratio schedule
-Common in natural environmental
-2 common variations–1) Random-Ratio, 2) progressive ratio

38
Q

1) random ratio

A

-Scheduling is controlled through a random number generator
-Produces similarly high rates of responding
-eg) casinos or video games

39
Q

2) Progressive ratio

A

-Ratio requirements move from small to large
-PRP increase with ratio size
-creates a “break point” measure of how hard an organism will work

40
Q

Fixed-Interval schedule

A

-Behavior is reinforced when it occurs after a given period of time
-Produce PRPs
-Responding increases gradually creating a scallop shape
-Uncommon in natural environment

41
Q

Variable interval schedule

A

-The timing of the response needed varies each time
-Interval varies around an average
-PRP’s are short and rare
-steady rates of responding, not as high as VR
-Common in natural environments
-

42
Q

Premack principle

A

-In nature different behaviors have different probabilities of occurring
-Low to high..reinforces low
-High to low.. does not reinforce high
-Any high probability response will can serve as a reinforcer for a lower probability response

43
Q

How to test premack principle

A

1) establish baseline responding for different behaviors
2) Instrumental conditioning procedure with low to high and high to low

44
Q

Example of premack principle

A

-If a child prefers playing pinball to eating veggies you can reinforce eating veggies by letting them play pinball each time they eat veggies
-high-probability behavior reinforces low-probability behavior

45
Q

Problems with premack principle

A

-Does not nicely account for conditional reinforcement
-low probability behavior can reinforce high-probability behavior when the organism has been deprived of the low probability behavior

46
Q

Antecedents/controlling stimuli

A

-Controlling stimulus is a stimulus that changes the probability of an operant behavior
-2 types
1) Discriminative stimulus
2) Extinction stimulus/ S delta

47
Q

1) Discriminative stimulus/occasion setter

A

-A stimulus or event that precedes an operant and sets the occasion for its reinforcement
-Makes behavior less likely to occur in the moment

48
Q

2) Extinction stimulus

A

-A stimulus or event that precedes an operant and sets the occasion for non-reinforcement
-Makes behavior less likely to occur in the moment

49
Q

Antecedents

A

-Include establishing and abolishing operations as well as control stimuli
-Evoke a behavior
-Alter the current probability of behavior

50
Q

Consequences

A

-Include reinforcers and punishers
-Strengthen or weaken the behavior
-Alter they future probability of behavior

51
Q

When does discrimination occur?

A

-When the presence or absence of stimuli is the occasion on which a response will be followed by reinforcement
eg) Pecking is only reinforced when the green light is on
-The green light IS the occasion when pecking will be reinforced

52
Q

What does discrimination refer to?

A

-The effect an occasion setting contingency has on behavior
-Refers to the effect of the response being more likely to occur in the presence of the SD than its absence

53
Q

Stimulus control

A

A change in operant behavior that occurs when either S^D or S-Delta is presented

54
Q

Discrimination index (ID)/Discrimination ratio

A

A measure of the stimulus control exerted by an S^D or S-Delta

55
Q

Generalization

A

-Less precise control
-Obtained by training in a wide array of settings/stimuli

56
Q

Stimulus generalization

A

1) Process where once a CS has been established, similar stimuli may also produce a CR
2) process by which an operant response occurs to one discriminative stimulus, it also occurs to other similar stimuli

57
Q

Stimulus discrimination

A

1) process where we exhibit less pronounced CR to CSs that differ from the original CS
2) Process where less responding occurs to stimuli that are different from the original trained stimulus
-Operant response to trained stimuli but not others

58
Q

Concept formation

A

1) The generalization within classes of stimuli; and
2) The discrimination between classes of stimuli

59
Q

Generalization in practice

A

-Occurred when the target behavior occurs in situations other than the specific training conditions
-Ideally it involves having the target behavior occur in all relevant situations

60
Q

Promoting generalization

A

-Reinforce occurrences of generalization
-Training setting and criterion setting should be quite similar
-Training and criterion should gradually become more dissimilar

61
Q

Stimulus exemplar

A

Stimuli that represent the range of relevant stimulus situations in which the response should occur after training

62
Q

General case programming

A

Different S^D’s may require different responses to obtain the same reinforcer