This cluster shows how animals and people split their time between two jobs when the pay is different. It tells us that just counting rewards is not enough; timing, effort, and surprise also change what we pick. BCBAs can use these facts to set up token boards, work breaks, or reinforcer menus so clients stay happy and get more done.
When two activities offer different rewards, people and animals do not pick one and stick with it forever. Instead, they split their time in a way that roughly matches the rewards each option gives. If one activity pays off 70% of the time, a learner will spend about 70% of their effort there. This pattern is called the matching law, and it holds across species, settings, and reinforcer types.
The timing of rewards matters as much as how often they come. Research shows that earlier rewards pull behavior more strongly than later ones, even if the total payout is the same. This is why a learner might choose a small, immediate token over a larger but delayed reward. When you design token boards or reinforcer menus, the gap between earning and redeeming tokens shapes what clients will work hardest for.
Reinforcement history also plays a role. Choice does not update instantly when schedules change. Studies show that behavior shifts gradually after a ratio reversal, and past experience with a schedule continues to influence choices for some time. This means that if you change a client's reinforcement plan, expect a lag before their behavior fully reflects the new contingencies.
Variability itself can be reinforcing. Some learners prefer options that offer unpredictable rewards over ones that are fully predictable, even when the average payoff is the same. Variable-ratio token exchange schedules maintain higher response rates than fixed-ratio schedules. Knowing this helps you design token economies that keep engagement high across the entire session.
Common questions from BCBAs and RBTs
The matching law says that behavior allocation matches reinforcement allocation. If one behavior gets 60% of available reinforcement, a learner will do it about 60% of the time.
The problem behavior is likely paying off more often, faster, or with less effort. Use the matching law as a guide: increase the rate, immediacy, or quality of reinforcement for the replacement behavior until it clearly outcompetes the problem behavior.
Research shows that reinforcer size does not change how quickly behavior adjusts to rate differences. Rate and immediacy are stronger drivers of choice than magnitude alone.
Yes. Variable-ratio exchange schedules — where the number of tokens needed to redeem varies unpredictably — maintain higher response rates than fixed-ratio schedules, especially as exchange requirements get larger.
Not immediately. Research shows choice shifts gradually after a schedule change, with recent history still influencing behavior for some time. Plan for a transition period and track data daily when modifying schedules.