Applied Behavior Analysis & Conditioning Flashcards
The length of time from a specific cue to the onset of behavior is the:
A) Interval
B) Latency
C) Frequency
D) Topography
B) Latency
What is the number of times a response occurs during an observational period?
A) Rate
B) Duration
C) Intensity
D) Frequency
D) Frequency
If a behavior occurs during 25 out of 200 sample intervals, what is the score for this one-zero sampling?
A) 0.125
B) 1.25
C) 12.5
D) 175
A) 0.125
What is one-zero sampling?
A method of recording if a behavior occurs in a specific time interval.
In X seconds, did the behavior happen or not happen? One if yes, or zero if no.
Which of the following involves the systematic manipulation of environmental contingencies?
A) Functional analysis
B) Behavioral accountability
C) Statistical function analysis
D) Contingent assessment tool
A) Functional analysis
Which of these is NOT a factor in affecting the aversiveness of a trigger?
A) Distance
B) Orientation
C) Animation
D) None of the Above
D) None of the Above
What is a Response Rate aka Rate of Responding?
How many times a behavior occurs within a specific unit of time.
The study of the functional relations between behavior and environmental events is
Behavioral Analysis
The definition of Behavioral Analysis is
The study of the functional relations between behavior and environmental events.
A Functional Relation is
The tendency of one event to vary in a regular way with one or more other events
“If X, then Y” - more accurately X and Y vary together in some regular way.
Example: I turn the key and my car starts. I pick up a leash and my dog excited for a walk.
Overt Behavior is
Behavior that can be observed by someone other than the person performing it.
Covert Behavior is
Cannot be measured/observed by a third party.
Behavior that can only be observed by the person performing it.
Example: Thoughts
Respondent Behavior is
Reflexive Behavior
Behavior that is most readily influenced by events that precede it.
Examples: Startle responses, Blinking your eye when a puff of air hits it, emotional physiological responses.
Operant Behavior is
Behavior that is influenced by events that follow it.
In it’s simplest form behavior modification/support looks at behavior from two directions
NOT ENOUGH BEHAVIOR
Goal: Increase Frequency
Antecedent: Adding a Cue
Consequence: Reinforcement
TOO MUCH BEHAVIOR
Goal: Decrease Frequency
Antecedent: Deleting a Cue
Consequence: Punishment
Source: Raising Canine
Setting Events are
Antecedents
The conditions in the environment related to when the behavior appears.
Internal: Wellness
External: Location, Stimuli, Cues
Establishing Operations in an Antecedent refers to:
Concept of filling or draining the “gas tank” to alter the value of the reinforcer that you’ll use.
Example: Starting with your dog slightly hungry to increase the value of treats
Example: Being very boring when you come in the door so you are less rewarding
What is the defining feature of Primary Reinforcers?
A) They acquire their reinforcing properties through learning
B) They are dependent on a history of pairing with other reinforcement
C) They are never dependent on learning
D) They are usually social in nature
C) They are never dependent on learning
Behavior is the result of
Antecedents - that set the stage for it
Consequences - the history of what has happened in the past that strengthened this action
We can change these two values
When doing a Functional Assessment you should start by analyzing the:
A) Antecedent
B) Behavior
C) Consequence
D) Distant Antecedents
B) Behavior
The Reification Fallacy is
An explanatory fiction.
When an abstract concept or idea is treated as a concrete thing.
Examples: My dog wants to be the leader of the pack. My dog is jealous when I talk on the phone. My dog is a Super Male.
Also known as Hypostatization
The problems with labels or constructs are that
They cannot be tested or observed.
A Behavioral Repertoire refers to
Everything the animal (or person) knows how to do at a particular point in time
Learning History is
The history of exposure to antecedents and consequences
Bridge
An auditory, visual or haptic cue given immediately before the reward. Serving to bridge the gap between behavior and reward.
Common example: Clicker
Source: mentioned with no definition in Fight! By Jean Donaldson. Definition from Simon Pris
KGS
Keep Going Signal
A prompt or cue used to encourage a dog to continue performing a desired behavior, often used in behavior chains when transitioning between tasks
Different from a bridge/marker as there is no immediate reward, rather this is a prompt to continue an action.
Jean Donaldson
Joseph Wolpe
Coined Systematic Desensitization as an exposure therapy
Response Cost
A form of P-
The process of reducing the frequency of a target behavior by making removal of a reinforcer contingent on the target behavior
Example - Your dog is on a continuous reward schedule for walking on a loose leash, when they pull you do not give the treat
A Contrived Reinforcer is
A reinforcer that has been arranged for the purpose of modifying behavior.
Examples: Treats, toys, access to play
Versus
Natural Reinforcers which have not been arranged for the purpose of modifying behavior, spontaneous, unplanned. Example, chasing a bunny, an off leash dog coming to say hi.
CRF
Continuous Reinforcement Schedule
Every occurrence of the behavior is reinforced.
PRF
Partial or Intermittent Reinforcement Schedule
Behavior is reinforced only after certain responses
The four types of PRF are
Fixed Ratio (set number)
Fixed Interval (set time)
Variable Ratio (variable number)
Variable Interval (variable time)
FR-5 is an example of a PFR that means
Reward given after 5 occurrences.
Fixed Ratio, 5 behaviors gets a treat
A FI-5 PRF means
Reward given after 5 seconds.
Fixed Interval, 5 seconds of continuous behavior gets a treat. Example: Sit-Stay for 5 seconds gets a treat, no treat if pup gets up at 2 seconds
A VR-5 PRF means
Reward schedule varies. After 4, then 5, then 3, then 5, then 6 occurrences.
Variable Ratio with 5 being the average you are ping ponging around
A VI-5 PRF means
Reward schedule varies. After 4 seconds, then 5 seconds, then 3 seconds then 5 seconds then 6 swconds.
Variable Interval with 5 being the average length of TIME you are ping ponging around
Stimulus Discrimination
Tendency for behavior to occur at different frequencies in different situations
Discriminative Stimulus
Any stimulus (Antecedent) that produces behavior in a varying rate.
Example: Fido only jumps on Bob, not Jane because Jane turns away when Fido jumps but Bob always hugs Fido when he jumps. Fido has learned to discriminate between Bob and Jane.
Fido only jumps on Bob, not Jane because Jane turns away when Fido jumps but Bob always hugs Fido when he jumps. Fido has learned to discriminate between Bob and Jane.
Bob is the: ___
Jane is the: ___
Bob: SD
Jane: S^
SD: Discriminative Stimulus. Stimulus where behavior is reinforced
S∆: Stimulus Delta. Stimulus where behavior is not reinforced
Simultaneous Discrimination Training
Present two stimuli at the same time.
Example: Show a ball and a bone. Say “ball”. If the dog touches the ball, reward. If not, no reward.
Successive Discrimination Training
Present one stimulus after the other.
Example: After the dog knows the word “Ball” means touch the ball, start training the name for “Bone”
Frequency
The number of times a response occurs during an observational period.
Respondent Conditioning
Another name for Classical Conditioning or Pavlovian Conditioning.
Pairing a previously Neutral Stimulus with an Unconditioned Stimulus