Assessment & Research

Applying signal-detection theory to the study of observer accuracy and bias in behavioral assessment.

Lerman et al. (2010) · Journal of applied behavior analysis 2010
★ The Verdict

Observer scoring drifts when feedback or payoffs change—guard your data with checks and rotation.

✓ Read this if BCBAs who train staff to collect behavioral data in any setting.
✗ Skip if Practitioners using fully automated sensors with no human coder.

01Research in Context

01

What this study did

Chou et al. (2010) used signal-detection theory to test what changes observer accuracy. Adults without disabilities watched short clips and scored behaviors. The team gave different feedback, rewards, or clear definitions to see who would record correctly and who would drift.

02

What they found

Feedback and reward pushed observers to over- or under-count. Clear definitions helped a little, but money and praise shaped the numbers far more.

03

How this fits with other research

Meltzer (1983) showed the same pattern in pigeons: change the grain payoff and the birds’ bias swings. C et al. moved the idea to human staff.

Yaw et al. (2014) took the next step. They gave real residential staff feedback after training. Accuracy jumped, proving the lab finding works on the floor.

DeRoma et al. (2004) adds a twist. Their observers improved their own safety habits while watching others. C et al. focused on data drift; M et al. show watching can also change the watcher’s behavior.

Virues-Ortega et al. (2022) compared paper and app scoring. All tools were fairly accurate, so the tool matters less than the feedback the observer receives.

04

Why it matters

Your graph is only as good as the person holding the clicker. Schedule random reliability checks. Rotate data collectors. Give brief, neutral feedback right after the session. These small moves guard the numbers that guide your treatment decisions.

Free CEUs

Want CEUs on This Topic?

The ABA Clubhouse has 60+ free CEUs — live every Wednesday. Ethics, supervision & clinical topics.

Join Free →
→ Action — try this Monday

Pick one client, have a second staff silently collect one session, compare sheets, and give neutral feedback within five minutes.

02At a glance

Intervention
not applicable
Design
other
Sample size
61
Population
neurotypical
Finding
positive

03Original abstract

We evaluated the feasibility and utility of a laboratory model for examining observer accuracy within the framework of signal-detection theory (SDT). Sixty-one individuals collected data on aggression while viewing videotaped segments of simulated teacher-child interactions. The purpose of Experiment 1 was to determine if brief feedback and contingencies for scoring accurately would bias responding reliably. Experiment 2 focused on one variable (specificity of the operational definition) that we hypothesized might decrease the likelihood of bias. The effects of social consequences and information about expected behavior change were examined in Experiment 3. Results indicated that feedback and contingencies reliably biased responding and that the clarity of the definition only moderately affected this outcome.

Journal of applied behavior analysis, 2010 · doi:10.1901/jaba.2010.43-195