Reading Notes 5 Flashcards

1
Q

What two operations best characterized Thorndike’s experiments with cats?

A
  1. Establishing operations
  2. Consequential
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What did Thorndike call his experiments with his cats? Why?

A

Trial and error learning

  • It took the cats several attempts, with several errors, before they performed consistently and quickly
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Fill in the blank: Thorndike’s Law of Effect suggested that responses could be made ___________ probable by some consequences and __________ probable by others.

A

More
Less

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What do we call events that make a response more probable?

A

Reinforcers

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What do we call events that make responses less probable?

A

Punishers

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Maze started as very complicated constructions. They became increasingly simpler and then, were __________ as an apparatus for the study of nonhuman animal behavior

A

Dropped

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Experimental analysis of behavior (EAB) typically involves rats and ________ as subjects of study.

A

Pigeons

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

EAB also involves research with humans. But with humans, the apparatus is different. How is the apparatus different with humans?

A

Rats- press levers
Humans- press buttons, or keys on a keyboard

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Operant behavior is:

A

1) EMITTED (not elicited)
- Term used to describe behavior that “operates” on the environment

2) SELECTED by its consequences
- Recall Thorndike’s Law o/ Effect

3) MODIFIABLE by its consequences
- We change our behavior, based on consequences

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Operant

A

A response that produces a change in the environment and increases in frequency due to that change

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

T/F: No two responses are exactly alike

A

True

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

The ___________ of the response is less important than its function

A

Topography

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What are the different types of operant behavior?

A

1) Free operant
2) Restricted operant

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What is a free operant?

A

An org. may repeatedly engage in the response over an extended period o/ time

  • Key: The response can be made w/o any interference from the experimenter
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Give examples of free operants.

A

a. Talking to yourself (e.g. thinking)
b. Rat pressing a lever

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What is a restricted operant?

A

Rate o/ responding is determined by opportunities to emit the response

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

Give examples of restricted operants.

A

a. Thorndike’s puzzle box
b. Responding to flashcards

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

A high response rate results in what kind of slope?

A

Steep slope

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

A medium response rate results in what kind of slope?

A

Medium slope

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

A low response rate results in what kind of slope?

A

Small slope

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

How do you talk about behavior change?

A

1) One cannot reinforce a person
2) One cannot deliver reinforcement
3) By definition, a reinforcer results in an increase in the future likelihood o/ a response

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

“One cannot reinforce a person.” What does this mean?

A
  • Increasing the future p o/ that person
  • Instead, we say that we reinforce BEHAVIOR
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

“One cannot deliver reinforcement.” What does that mean?

A
  • We only deliver REINFORCERS
  • Reinforcement happens as a result o/ delivering reinforcing
  • Delivering reinforcers = operation (IV, what we did/the cause)
  • Reinforcement or the increase in behavior = the process (DV, what happened/the effect)
24
Q

“By definition, a reinforcer results in an increase in the future likelihood o/ a response.” What does this mean?

A
  • One can’t say that reinforcement didn’t work, bc he didn’t do his laundry
  • Why?- A reinforcer must increase a behavior. If it doesn’t increase behavior, it is NOT a reinforcer
25
Q

T/F: When a response is reinforced, it increases in frequency.

A

True

26
Q

T/F: When a response is no longer reinforced, it decreases in frequency.

A

True

27
Q

What is the difference between extinction and process?

A

Extinction as operation (IV, what we do)- discontinuing reinforcer presentation

Process (DV, what happens)- when responding decreases to pre-reinforcement conditions, or is eliminated

28
Q

What are the side effects of extinction?

A

1) Extinction burst- relative increase in the rate o/ behavior
- Think o/ putting money in vending machine but nothing comes out

2) Greater variability in responding
- e.g. pushing other buttons

3) Increase in the force o/ the response
- e.g. push buttons harder

4) Emotional responses
- eg. extinction-induced aggression

29
Q

What does EXT mean?

A

Withholding reinforcers for previously reinforced responses
- Behavior should stop occurring

30
Q

What happens during scenario 1 (CRF)

A

CRF = Continuous

  1. Deliver 1 food pellet for every lever press for 60 days
  2. Implement EXT
31
Q

What happens during scenario 2 (INT)?

A

INT = Intermittent

  1. Deliver 1 food pellet for every 10 or so responses, for 60 days
  2. Implement EXT
32
Q

T/F: Behavior under a CRF schedule is easier to extinguish when compared to intermittent schedules. Explain.

A

True

Explanation- If every R is reinforced (CRF) it’s easier to tell when reinforcers stop being delivered

33
Q

What is discrimination learning?

A

Process o/ being able to tell the diff. between situation

34
Q

What is spontaneous recovery?

A

An increase in the rate o/ responding (above operant levels) after and during exposure to EXT

35
Q

What is the main difference between spontaneous recovery and resurgence?

A

Resurgence involves 1+ behavior

36
Q

Name two examples of the recovery of extinguished behavior.

A
  1. Spontaneous recovery
  2. Resurgence
37
Q

Explain what occurs during EXT, when exposed to the next session, and at the end.

A

During EXT- responding decreases across the session (over time)
- You respond less at the end than at the beginning

Exposed to the next session- you don’t respond as low as you did at the end o/ the period EXT session

At the end- you respond at slightly higher levels
- Behavior decreases during the course o/ the session

38
Q

What is a good way to think of your behavior? Explain the analogy.

A

Power outage
a. You pick up the phone to call a friend, but your behavior is met w/ EXT- phone is dead
b. Instead, you decide to read your textbook by a candle
c. The power is still off, hours later. You get hungry and decide to order a pizza and pick up the phone
d. You realize the power’s out

39
Q

What is resurgence?

A

Return o/ previously established behavior during EXT o/ more recently reinforced behavior

40
Q

Give an example of resurgence. Explain.

A

A rat can: (a) emit a lever (b) pull a chain
1. When we reinforce lever presses, the response increases in frequency

  1. Chain pulls are reinforced, so now that response occurs frequently
    - Lever presses aren’t reinforced so that response doesn’t occur
  2. In the last phase o/ the experiment, we extinguish BOTH responses (resurgence condition)
  3. We expect to see the lever press (it worked first)

Lever press and chain pulls —> aggression reinforced by parent attention and appropriate requests for attention

41
Q

What is the Premack Principle?

A

States that a higher frequency (probability) response will reinforce a lower frequency (probability) response

42
Q

Name a few examples of how the Premack Principle is applied in real life.

A
  1. High probability response = watching your favorite show
    - Low probability response = studying
  2. Applying the principle —> Memorize 5 definitions THEN you can watch one episode o/ your favorite show
  3. High probability response = eating gummy bears
    - Low probability response = reading a chapter
  4. Applying the principle —> Read one paragraph THEN you can have a gummy bear
43
Q

What are conditioned reformers? Give examples.

A

Becomes effective as a reinforcer through its relation to some other reinforcer

  • e.g. High five, good grades, bell in operant behavior
44
Q

What is a generalized conditioned reinforcer? Give examples.

A

Becomes effective as a reinforcer through its relation to many other reinforcers

e.g. Money, gift cards

45
Q

What is an unconditioned response? Give examples.

A

Does not depend on a relation to other to other reinforcers

e.g. Food, water, sleep

46
Q

What is an automatic reinforcer? Give examples?

A

One that has a natural relation to the response that produces it (aka, intrinsic reinforcer)

e.g. Thumb sucking, hair twirling, & nail-biting

47
Q

What is a contrived reinforcer? Give examples

A

One that has an arbitrary relation to the responses (aka, extrinsic reinforcer)

e.g. Paycheck, chocolate

48
Q

What is a conditioned reinforcer?

A

Becomes effective as a reinforcer through its RELATION to some other reinforcer

49
Q

Give examples of conditioned reinforcers.

A

High fives, good grades, bell in operant behavior

50
Q

What is a generalized conditioned reinforcer?

A

Becomes effective as a reinforcer through its relation to MANY other reinforcers

51
Q

Give examples of generalized conditioned reinforcers.

A

Money, gift cards

52
Q

What is an unconditioned reinforcer?

A

Doesn’t depend on a relation to other reinforcers

53
Q

Give examples of unconditioned reinforcers

A

Food, water, & sleep

54
Q

What is an automatic reinforcer?

A

One that has a natural relation to the response that produces it (aka, intrinsic reinforcer)

55
Q

Give examples of automatic reinforcers

A

Thumb sucking, hair twirling, & nail biting

56
Q

What are contrived reinforcers?

A

One that has an arbitrary relation to the responses (aka, extrinsic reinforcer)

57
Q

Give examples of contrived reinforcers.

A

Paycheck, chocolate