Observing responses and serial stimuli: searching for the reinforcing properties of the S-.
A no-reward signal can turn into a tiny reward if it shows up late in a long extinction stretch.
01Research in Context
What this study did
Escobar et al. (2009) worked with pigeons in a lab. They set up a mixed schedule where food sometimes came after a green light and never came after a red light.
The red light stayed on for a long stretch. Near the end of that stretch, the red light blinked in a set pattern. The birds could peck a small key just to see which light was on.
The team asked: Do those late, serial red blinks make the birds peek more? If yes, the red blinks act like a tiny reward.
What they found
Peeks at the lights rose only when the serial red blinks showed up late in the no-food period.
The red light, once just a signal for no food, turned into a conditioned reinforcer for the observing response.
How this fits with other research
Odom et al. (1986) saw more observing any time a red light popped up, even when it gave no extra reward. They warned that simple S- presentation can fool you into thinking the signal is reinforcing.
Rogelio’s team took that warning and added a timing twist. By placing the serial S- late in extinction, they showed the same stimulus can switch from neutral to reinforcing.
Bland et al. (2018) extend the idea in the other direction: when S- is given right after a response, it can punish that response. One stimulus, two jobs—reinforcer or punisher—depending on when and how you deliver it.
Why it matters
Your client may keep looking at, asking about, or even approaching items linked with "no" if those cues pop up after a long wait. Try spacing those cues and watch if the behavior grows. If it does, the cue itself may be feeding the behavior. Shift timing, add prompts, or reinforce alternate looks to keep the session on track.
Want CEUs on This Topic?
The ABA Clubhouse has 60+ free CEUs — live every Wednesday. Ethics, supervision & clinical topics.
Join Free →During a long DRO interval, remove or vary the "no" signal near the end to block accidental reinforcement of looking.
02At a glance
03Original abstract
The control exerted by a stimulus associated with an extinction component (S-) on observing responses was determined as a function of its temporal relation with the onset of the reinforcement component. Lever pressing by rats was reinforced on a mixed random-interval extinction schedule. Each press on a second lever produced stimuli associated with the component of the schedule in effect. In Experiment 1 a response-dependent clock procedure that incorporated different stimuli associated with an extinction component of a variable duration was used. When a single S- was presented throughout the extinction component, the rate of observing remained relatively constant across this component. In the response-dependent clock procedure, observing responses increased from the beginning to the end of the extinction component. This result was replicated in Experiment 2, using a similar clock procedure but keeping the number of stimuli per extinction component constant. We conclude that the S- can function as a conditioned reinforcer, a neutral stimulus or as an aversive stimulus, depending on its temporal location within the extinction component.
Journal of the experimental analysis of behavior, 2009 · doi:10.1901/jeab.2009.92-215