Applying signal-detection theory to the study of observer accuracy and bias in behavioral assessment.
Observer scoring drifts when feedback or payoffs change—guard your data with checks and rotation.
01Research in Context
What this study did
Chou et al. (2010) used signal-detection theory to test what changes observer accuracy. Adults without disabilities watched short clips and scored behaviors. The team gave different feedback, rewards, or clear definitions to see who would record correctly and who would drift.
What they found
Feedback and reward pushed observers to over- or under-count. Clear definitions helped a little, but money and praise shaped the numbers far more.
How this fits with other research
Meltzer (1983) showed the same pattern in pigeons: change the grain payoff and the birds’ bias swings. C et al. moved the idea to human staff.
Yaw et al. (2014) took the next step. They gave real residential staff feedback after training. Accuracy jumped, proving the lab finding works on the floor.
DeRoma et al. (2004) adds a twist. Their observers improved their own safety habits while watching others. C et al. focused on data drift; M et al. show watching can also change the watcher’s behavior.
Virues-Ortega et al. (2022) compared paper and app scoring. All tools were fairly accurate, so the tool matters less than the feedback the observer receives.
Why it matters
Your graph is only as good as the person holding the clicker. Schedule random reliability checks. Rotate data collectors. Give brief, neutral feedback right after the session. These small moves guard the numbers that guide your treatment decisions.
Want CEUs on This Topic?
The ABA Clubhouse has 60+ free CEUs — live every Wednesday. Ethics, supervision & clinical topics.
Join Free →Pick one client, have a second staff silently collect one session, compare sheets, and give neutral feedback within five minutes.
02At a glance
03Original abstract
We evaluated the feasibility and utility of a laboratory model for examining observer accuracy within the framework of signal-detection theory (SDT). Sixty-one individuals collected data on aggression while viewing videotaped segments of simulated teacher-child interactions. The purpose of Experiment 1 was to determine if brief feedback and contingencies for scoring accurately would bias responding reliably. Experiment 2 focused on one variable (specificity of the operational definition) that we hypothesized might decrease the likelihood of bias. The effects of social consequences and information about expected behavior change were examined in Experiment 3. Results indicated that feedback and contingencies reliably biased responding and that the clarity of the definition only moderately affected this outcome.
Journal of applied behavior analysis, 2010 · doi:10.1901/jaba.2010.43-195