Exam 3 (part two) Flashcards
(19 cards)
Schedules of reinforcement (or punishment)
– schedule of consequence; the pattern in which we deliver reinforcement
Contingent schedules
– the presentation of a consequence depends on a response; a response is needed to produce reinforcement
Contingent schedules
- Continuous schedule – every response produces a reinforcer (FR1)
Contingent schedules
- Intermittent schedule – not every response produces a reinforcer
- Much more common in daily life
Fixed ratio (FR)
– the same number of responses every trial is required to produce reinforcement
- Ratio schedule – based on number of responses
- FR5 = 5 responses needed each trial
- Produces a break-and-run pattern
- Examples:
- Sales and commission
- Reading # of pages
- Buy 5, get 1 free
- Pair discussions
Post-reinforcement paus (PRP)
– break between reinforcements
- Positive relation between size of schedule and length of break
- More responses = longer break
Fixed interval (FI)
the same period of time produces reinforcement
- First response after interval of time produces reinforcement
- 30”, 30”, 30”, 30” …
- 30 seconds is when the reinforcement becomes AVAILABLE, not when it occurs
- Produced a scalloped pattern
- One response is required for reinforcement to occur
- Responding early is not reinforced; getting closer to the end of the interval increases the number of responses
- Examples:
- Washing machine
- Cooking food
- Few real-life examples
- FI schedules – why study?
- Parsimony – start by understanding the simplest thing and build up to complex
- Serve as a baseline for understanding the effects of other
Variable ratio (VR)
– a different number of responses produce reinforcement
- VR15: 5, 20, 15, 30, 10
- 15 is the average number of responses is
- Produces a high, steady rate of response
- The faster you respond, the faster you get to reinforcement
- Examples:
- Gambling
- Asking for dates
Variable interval (VI)
– a different period of time produces reinforcement
- 30”, 15”, 60”, 10” …
- Produces a moderate, steady rate of response
- Moderate because responding faster doesn’t produce a reinforcement any faster, but you could miss the reinforcement if you respond too slow
- Examples:
- Checking for texts
- Fishing
- Pop quizzes
- Ratio schedules vs. interval schedules
- Ratio schedules produce higher rates of response
- VR is the fastest
- FR is second fastest; approximately equal to VI
- FI is the slowest
Variable interval (VI)
- Ratio schedules vs. interval schedules
- Ratio schedules produce higher rates of response
- VR is the fastest
- FR is second fastest; approximately equal to VI
- FI is the slowest
Noncontingent (response-independent) schedules
– you do not need to do anything to get reinforcement
- Often FT or VT schedules
- Responsible for superstitious behaviors
- “Learned laziness”
Fixed time
same period of time
Variable time
different period of time
Theory of reinforcement
– WHY does R increase?
Hull’s drive-reduction theory
- (“drive”) R ® SR
- Inside the organism is some unobserved drive
- Biological in nature (e.g., hunger, thirst, reproduction, sleep)
- R occurs because SR reduces physiological drive
- No longer widely accepted in psychology of learning
- Many things we do are not tied to a biological drive
- Hull did all of his research with nonhuman subjects
Premack principle
- Reinforcement – LPB (lower probability behavior) ® HPB (higher probability behavior)
- NOT low and high; lower and higher; reinforcers are relative
- LPB (response) produces opportunity for HPB (reinforcer)
- Lever press ® EATING food
- Doing homework ® WATCHING Netflix
- Doing homework reinforced by watching Netflix
- Watching TV (HPB) ® cleaning (LPB)
- Punishment
- You are punishing watching TV by following the HPB with the LPB
- R (HBP) stops because it produces punisher (LPB)
Premack principle
- Implications:
- Stimulus vs. behavior
- It is not the stimulus (food) producing reinforcement, but the opportunity to behave a certain way (eating)
- Primary vs. secondary
- No distinction
- Reversibility (no such thing as something that is ALWAYS a reinforcer)
- Applications:
- Order your day: LPBs to HPBs
Response deprivation hypothesis
- Behavior = reinforcers
- Reinforcement:
- (a) Behavior 1 (SR) restricted
- (b) Behavior 1 falls below its baseline (deprivation)
- If (a) and (b) are met, behavior 1 reinforces any other behavior
- R increases so behavior 1 (SR) can get back to baseline
- Ex with rat and lever:
- Pressing lever (1 min per day)
- By making it press the lever, we are restricting the amount of time it can spend eating
- Eating = 1 hour per day
- Pressing lever = eating falls below baseline
- Eating can reinforce lever pressing
- Pressing more = access to eating (back up to baseline)
- Ex with TV:
- TV watching (6 hours per day)
- Studying = less TV watched; falls below baseline
- TV watching can reinforce studying
- Studying more = access to TV (back up to 6 hours)
Premack vs. response deprivation hypothesis
- LPB (studying) ® HPB (watching TV below baseline)
- Premack = reinforcement
- Response deprivation hypothesis = reinforcement
- Increase in studying
- LPB (studying) ® HPB (watching TV above baseline)
- Premack = reinforcement
- Response deprivation hypothesis = not reinforcement
- Not below baseline
- HPB (watching TV) ® LPB (studying below baseline)
- Premack = not reinforcement (punishment)
- HPB follows LPB
- Response deprivation hypothesis = reinforcement
- Increase in watching TV
- Response deprivation hypothesis is more accurate theory; Premack is still useful