ABA Fundamentals

Effects of intertrial reinforcers on self-control choice.

Mazur (1994) · Journal of the experimental analysis of behavior 1994
★ The Verdict

Extra reinforcers delivered during delay intervals make organisms choose smaller, sooner rewards more often.

✓ Read this if BCBAs designing token economies, delay tolerance programs, or reinforcement schedules with wait components
✗ Skip if Practitioners working only with immediate reinforcement or non-delayed contingencies

01Research in Context

01

What this study did

The researchers set up a classic self-control game for pigeons. Birds could pick a small food reward right away or wait for a bigger reward later. During the wait time, the team slipped in extra bits of food on a variable schedule. They wanted to see if these free snacks would make the birds act more 'impulsive' and take the smaller, faster payoff.

The extra food came at different rates. Sometimes the birds got more free snacks, sometimes fewer. The scientists tracked how often each bird chose the quick, small reward versus the bigger, delayed one.

02

What they found

The free snacks worked like a charm. Birds chose the small, immediate reward more often when they got extra food during the wait. The effect grew stronger as the snack rate increased. The results fit the hyperbolic decay model - each extra snack made the delayed reward lose value faster.

In plain terms: more free food now equals less patience for the bigger payoff later.

03

How this fits with other research

This finding flips an earlier result on its head. Higgins et al. (1992) found that fewer trials (lower 'income') made rats more patient. Wulfert (1994) shows more free food during trials makes pigeons less patient. The difference? T's team reduced the total number of earning opportunities. E's team added extra, unearned rewards during waits. Both changes affect how organisms value delayed payoffs, but in opposite directions.

The study extends work by Clarke et al. (2003) and Carlin et al. (2012). Those teams taught humans with intellectual disabilities to wait for bigger rewards by giving them activities during delays. Wulfert (1994) shows that what you do during the wait matters - free food rewards increase impulsivity, while engaging activities can build patience.

The results support the hyperbolic decay model first tested by Sanford et al. (1980). That study showed how absolute delay changes preference. Wulfert (1994) adds that the density of rewards during those delays also shapes choice, consistent with the same mathematical model.

04

Why it matters

When you program delays in reinforcement schedules, watch what happens during the wait. Giving clients extra tokens, snacks, or attention while they wait for a bigger reinforcer might accidentally teach them to pick smaller, faster payoffs instead. If you want to build patience, keep the delay period lean. Save the bonus reinforcers for after the client chooses the larger, delayed option. This is especially important in token economies, DRO procedures, or any intervention where clients wait for better rewards.

Free CEUs

Want CEUs on This Topic?

The ABA Clubhouse has 60+ free CEUs — live every Wednesday. Ethics, supervision & clinical topics.

Join Free →
→ Action — try this Monday

Remove any unearned snacks, tokens, or attention during delay periods in your reinforcement schedule

02At a glance

Intervention
other
Design
single case other
Population
not specified
Finding
positive

03Original abstract

In three experiments, pigeons chose between a small amount of food delivered after a short delay and a larger amount delivered after a longer delay. A discrete-trial adjusting-delay procedure was used to estimate indifference points--pairs of delay-amount combinations that were chosen about equally often. In Experiment 1, when additional reinforcers were available during intertrial intervals on a variable-interval schedule, preference for the smaller, more immediate reinforcer increased. Experiment 2 found that this shift in preference occurred partly because the variable-interval schedule started sooner after the smaller, more immediate reinforcer, but there was still a small shift in preference when the durations and temporal locations of the variable-interval schedules were identical for both alternatives. Experiment 3 found greater increases in preference for the smaller, more immediate reinforcer with a variable-interval 15-s schedule than with a variable-interval 90-s schedule. The results were generally consistent with a model that states that the impact of any event that follows a choice response declines according to a hyperbolic function with increasing time since the moment of choice.

Journal of the experimental analysis of behavior, 1994 · doi:10.1901/jeab.1994.61-83