Reading Notes 5 Flashcards

1
Q

What two operations best characterized Thorndike’s experiments with cats?

A
  1. Establishing operations
  2. Consequential
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What did Thorndike call his experiments with his cats? Why?

A

Trial and error learning

  • It took the cats several attempts, with several errors, before they performed consistently and quickly
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Fill in the blank: Thorndike’s Law of Effect suggested that responses could be made ___________ probable by some consequences and __________ probable by others.

A

More
Less

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What do we call events that make a response more probable?

A

Reinforcers

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What do we call events that make responses less probable?

A

Punishers

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Maze started as very complicated constructions. They became increasingly simpler and then, were __________ as an apparatus for the study of nonhuman animal behavior

A

Dropped

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Experimental analysis of behavior (EAB) typically involves rats and ________ as subjects of study.

A

Pigeons

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

EAB also involves research with humans. But with humans, the apparatus is different. How is the apparatus different with humans?

A

Rats- press levers
Humans- press buttons, or keys on a keyboard

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Operant behavior is:

A

1) EMITTED (not elicited)
- Term used to describe behavior that “operates” on the environment

2) SELECTED by its consequences
- Recall Thorndike’s Law o/ Effect

3) MODIFIABLE by its consequences
- We change our behavior, based on consequences

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Operant

A

A response that produces a change in the environment and increases in frequency due to that change

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

T/F: No two responses are exactly alike

A

True

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

The ___________ of the response is less important than its function

A

Topography

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What are the different types of operant behavior?

A

1) Free operant
2) Restricted operant

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What is a free operant?

A

An org. may repeatedly engage in the response over an extended period o/ time

  • Key: The response can be made w/o any interference from the experimenter
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Give examples of free operants.

A

a. Talking to yourself (e.g. thinking)
b. Rat pressing a lever

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What is a restricted operant?

A

Rate o/ responding is determined by opportunities to emit the response

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

Give examples of restricted operants.

A

a. Thorndike’s puzzle box
b. Responding to flashcards

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

A high response rate results in what kind of slope?

A

Steep slope

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

A medium response rate results in what kind of slope?

A

Medium slope

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

A low response rate results in what kind of slope?

A

Small slope

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

How do you talk about behavior change?

A

1) One cannot reinforce a person
2) One cannot deliver reinforcement
3) By definition, a reinforcer results in an increase in the future likelihood o/ a response

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

“One cannot reinforce a person.” What does this mean?

A
  • Increasing the future p o/ that person
  • Instead, we say that we reinforce BEHAVIOR
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

“One cannot deliver reinforcement.” What does that mean?

A
  • We only deliver REINFORCERS
  • Reinforcement happens as a result o/ delivering reinforcing
  • Delivering reinforcers = operation (IV, what we did/the cause)
  • Reinforcement or the increase in behavior = the process (DV, what happened/the effect)
24
Q

“By definition, a reinforcer results in an increase in the future likelihood o/ a response.” What does this mean?

A
  • One can’t say that reinforcement didn’t work, bc he didn’t do his laundry
  • Why?- A reinforcer must increase a behavior. If it doesn’t increase behavior, it is NOT a reinforcer
25
T/F: When a response is reinforced, it increases in frequency.
True
26
T/F: When a response is no longer reinforced, it decreases in frequency.
True
27
What is the difference between extinction and process?
Extinction as operation (IV, what we do)- discontinuing reinforcer presentation Process (DV, what happens)- when responding decreases to pre-reinforcement conditions, or is eliminated
28
What are the side effects of extinction?
1) Extinction burst- relative increase in the rate o/ behavior - Think o/ putting money in vending machine but nothing comes out 2) Greater variability in responding - e.g. pushing other buttons 3) Increase in the force o/ the response - e.g. push buttons harder 4) Emotional responses - eg. extinction-induced aggression
29
What does EXT mean?
Withholding reinforcers for previously reinforced responses - Behavior should stop occurring
30
What happens during scenario 1 (CRF)
CRF = Continuous 1. Deliver 1 food pellet for every lever press for 60 days 2. Implement EXT
31
What happens during scenario 2 (INT)?
INT = Intermittent 1. Deliver 1 food pellet for every 10 or so responses, for 60 days 2. Implement EXT
32
T/F: Behavior under a CRF schedule is easier to extinguish when compared to intermittent schedules. Explain.
True Explanation- If every R is reinforced (CRF) it's easier to tell when reinforcers stop being delivered
33
What is discrimination learning?
Process o/ being able to tell the diff. between situation
34
What is spontaneous recovery?
An increase in the rate o/ responding (above operant levels) after and during exposure to EXT
35
What is the main difference between spontaneous recovery and resurgence?
Resurgence involves 1+ behavior
36
Name two examples of the recovery of extinguished behavior.
1. Spontaneous recovery 2. Resurgence
37
Explain what occurs during EXT, when exposed to the next session, and at the end.
During EXT- responding decreases across the session (over time) - You respond less at the end than at the beginning Exposed to the next session- you don't respond as low as you did at the end o/ the period EXT session At the end- you respond at slightly higher levels - Behavior decreases during the course o/ the session
38
What is a good way to think of your behavior? Explain the analogy.
Power outage a. You pick up the phone to call a friend, but your behavior is met w/ EXT- phone is dead b. Instead, you decide to read your textbook by a candle c. The power is still off, hours later. You get hungry and decide to order a pizza and pick up the phone d. You realize the power's out
39
What is resurgence?
Return o/ previously established behavior during EXT o/ more recently reinforced behavior
40
Give an example of resurgence. Explain.
A rat can: (a) emit a lever (b) pull a chain 1. When we reinforce lever presses, the response increases in frequency 2. Chain pulls are reinforced, so now that response occurs frequently - Lever presses aren't reinforced so that response doesn't occur 3. In the last phase o/ the experiment, we extinguish BOTH responses (resurgence condition) 4. We expect to see the lever press (it worked first) Lever press and chain pulls ---> aggression reinforced by parent attention and appropriate requests for attention
41
What is the Premack Principle?
States that a higher frequency (probability) response will reinforce a lower frequency (probability) response
42
Name a few examples of how the Premack Principle is applied in real life.
1. High probability response = watching your favorite show - Low probability response = studying 2. Applying the principle ---> Memorize 5 definitions THEN you can watch one episode o/ your favorite show 3. High probability response = eating gummy bears - Low probability response = reading a chapter 4. Applying the principle ---> Read one paragraph THEN you can have a gummy bear
43
What are conditioned reformers? Give examples.
Becomes effective as a reinforcer through its relation to some other reinforcer - e.g. High five, good grades, bell in operant behavior
44
What is a generalized conditioned reinforcer? Give examples.
Becomes effective as a reinforcer through its relation to many other reinforcers e.g. Money, gift cards
45
What is an unconditioned response? Give examples.
Does not depend on a relation to other to other reinforcers e.g. Food, water, sleep
46
What is an automatic reinforcer? Give examples?
One that has a natural relation to the response that produces it (aka, intrinsic reinforcer) e.g. Thumb sucking, hair twirling, & nail-biting
47
What is a contrived reinforcer? Give examples
One that has an arbitrary relation to the responses (aka, extrinsic reinforcer) e.g. Paycheck, chocolate
48
What is a conditioned reinforcer?
Becomes effective as a reinforcer through its RELATION to some other reinforcer
49
Give examples of conditioned reinforcers.
High fives, good grades, bell in operant behavior
50
What is a generalized conditioned reinforcer?
Becomes effective as a reinforcer through its relation to MANY other reinforcers
51
Give examples of generalized conditioned reinforcers.
Money, gift cards
52
What is an unconditioned reinforcer?
Doesn't depend on a relation to other reinforcers
53
Give examples of unconditioned reinforcers
Food, water, & sleep
54
What is an automatic reinforcer?
One that has a natural relation to the response that produces it (aka, intrinsic reinforcer)
55
Give examples of automatic reinforcers
Thumb sucking, hair twirling, & nail biting
56
What are contrived reinforcers?
One that has an arbitrary relation to the responses (aka, extrinsic reinforcer)
57
Give examples of contrived reinforcers.
Paycheck, chocolate