Preference pulses induced by reinforcement.
A single reinforcer sparks a short-lived burst of preference for the source, even after the pay stops.
01Research in Context
What this study did
Hachiga et al. (2014) watched rats choose between two levers. One lever paid off on a VR-20 schedule. The other lever sometimes stayed on VR-20 and sometimes went into extinction.
The team tracked which lever the rat pressed second-by-second. They wanted to see if a reinforcer creates a short-lived burst of preference for the lever that delivered it.
What they found
Right after a pellet dropped, rats kept pressing the same lever for a few seconds. This mini-preference faded faster when the lever entered extinction.
If the lever still paid on VR-20, the brief spike lasted a bit longer. The data show a clear post-reinforcer pulse, even when the lever no longer pays.
How this fits with other research
Landon et al. (2003) ran a similar rat-lever setup and found bigger pellets make bigger pulses. Together the two studies show both size and extinction matter to the same momentary effect.
Hunter et al. (2019) conceptually replicated the idea with humans and toys. One accidental reinforcer after extinction pulled kids back to a toy, matching the rat pattern.
Cohen et al. (1993) seems to disagree. They saw reinforcers lengthen the current stay but claimed no real preference shift. The clash is timing: L looked at whole-session totals while Yosuke zoomed in on seconds. Both can be true.
Why it matters
Your client may return to a task right after praise, then drift away. That brief surge is the preference pulse. Use it: deliver the reinforcer, then immediately prompt the next response while the pulse is alive. If the task is now harder or unreinforced, move fast—the pulse dies quicker under extinction.
Want CEUs on This Topic?
The ABA Clubhouse has 60+ free CEUs — live every Wednesday. Ethics, supervision & clinical topics.
Join Free →After you deliver a reinforcer, prompt the next response within two seconds to ride the preference pulse.
02At a glance
03Original abstract
Eight rats responded on concurrent Variable-Ratio 20 Extinction schedules for food reinforcement. The assignment of variable-ratio reinforcement to a left or right lever varied randomly following each reinforcer, and was cued by illumination of a stimulus light above that lever. Postreinforcement preference levels decreased substantially and reliably over time when the lever that just delivered reinforcement was now in extinction; however, if that lever was once again associated with variable ratio, this decrease in same-lever preference tended to be small, and for some subjects, not in evidence. The changes in preference level to the extinction lever were well described by a modified version of Killeen, Hanson, and Osborne's (1978) induction model. Consistent with this model's attribution of preference change to induction, we attribute preference change in this report to a brief period of reinforcer-induced arousal that energizes responding to the lever that delivered the last reinforcer. After a few seconds, this induced responding diminishes, and the operant responding that remains comes under the control of the stimulus light cuing the lever providing variable-ratio reinforcement.
Journal of the experimental analysis of behavior, 2014 · doi:10.1002/jeab.108