Preference for and effects of variable-as opposed to fixed-reinforcer duration.
Unpredictable reward size can beat a steady one even when it pays less.
01Research in Context
What this study did
Researchers let pigeons choose between two keys. One key gave food for the same length of time every time. The other key gave food for different lengths of time.
The birds could switch keys whenever they wanted. The team counted pecks on each key to see which schedule the birds liked more.
What they found
The pigeons pecked more on the variable key. They stuck with it even when it gave less total food.
The birds picked the gamble even when it paid less. The surprise of not knowing the next prize kept them playing.
How this fits with other research
English et al. (1995) ran the same test with water instead of food. The pigeons still chose the variable schedule. The preference holds across different kinds of reinforcers.
Andrade et al. (2014) showed you can thin human rewards to every few days and still keep step counts high. That study used fixed schedules, yet the birds here loved the opposite. The difference is purpose: F et al. wanted to save tokens; M et al. wanted to see what animals pick when given a choice.
Lovaas et al. (1973) spread news clips and tokens across short sessions. They focused on learning, not preference. Together the papers show timing matters, but the question you ask—learning vs liking—shapes the schedule you pick.
Why it matters
Your client may choose an uncertain reward over a sure one even if it gives less. Build in small surprises—mystery prizes, random bonus tokens—to keep engagement high. Just watch the net payoff; variability helps motivation, but you still need enough reinforcement to maintain the skill.
Want CEUs on This Topic?
The ABA Clubhouse has 60+ free CEUs — live every Wednesday. Ethics, supervision & clinical topics.
Join Free →Add one mystery reinforcer this week—wrap a token in foil so they don’t know if it’s 1 or 5 minutes of iPad—and track if responding stays strong.
02At a glance
03Original abstract
Pigeons were trained on multiple schedules in which a fixed number of pecks produced either a fixed or a variable period of access to food, the average variable-duration reinforcement equalling the fixed. Pecking rates were generally higher during the variable-duration component. Subsequent performance on concurrent schedules revealed an initial preference for variable-duration reinforcement for all subjects; for most subjects, this preference was sustained. For one subject, the average variable duration was gradually reduced to half the fixed duration: continued preference for the variable component resulted in a loss of up to 30% of available reinforcement time. A return to multiple schedules with unequal pay-off shifted the preference to the greater fixed duration, and this preference was maintained even when the variable duration was again raised to equal the fixed duration. For the remaining subjects, the initial variable-duration preference on concurrent schedules was gradually replaced by a side preference. When the range of variable durations was varied, keeping the average variable duration equal to the fixed, the occasional longer reinforcers sustained a preference for variable-reinforcer durations for three of the four subjects.
Journal of the experimental analysis of behavior, 1974 · doi:10.1901/jeab.1974.21-89