Assessment & Research

Comparing masked and traditional visual analysis of multiple‐baseline designs

Wolfe et al. (2026) · Journal of Applied Behavior Analysis 2026
★ The Verdict

Masked visual analysis agrees with traditional checking most of the time, so you can use it as a quick shield against false positives.

✓ Read this if BCBAs who run or review multiple-baseline studies in clinics or schools.
✗ Skip if Practitioners who already use automated decision tools for every graph.

01Research in Context

01

What this study did

Wolfe et al. (2026) asked experts to look at multiple-baseline graphs two ways. First, they did the usual visual check. Then they did the same check again, but the names of the phases were hidden.

The team counted how often raters agreed with themselves and with each other under each method.

02

What they found

Both masked and traditional checks gave moderate in-method agreement. On two-thirds of the graphs, the hidden-phase check reached the same yes-or-no decision as the open-label check at least 75 percent of the time.

In short, masking did not wreck reliability; it stayed in the same ballpark as the usual way.

03

How this fits with other research

Ferron et al. (2017) ran computer simulations and said masking keeps false alarms under five percent while still catching real effects. Wolfe et al. now show live experts behave the same way, so the simulation holds up in practice.

Lanovaz et al. (2021) found a machine-learning script beats naked-eye calls. Wolfe et al. do not contradict this; they simply show that if you stay with human eyes, adding a mask is a cheap safety step.

Adams et al. (2024) pushed computer scripts for functional-analysis graphs. Wolfe et al. keep the focus on plain visual rules, giving BCBAs a low-tech option before jumping to code.

04

Why it matters

You can add masked review to any multiple-baseline study without new software or stats classes. It takes five minutes to cover phase labels before you ask a colleague to check your graph. That small step guards against seeing an effect that is not there, keeping your single-case conclusions clean.

Free CEUs

Want CEUs on This Topic?

The ABA Clubhouse has 60+ free CEUs — live every Wednesday. Ethics, supervision & clinical topics.

Join Free →
→ Action — try this Monday

Before you show your next MB graph to a colleague, hide the phase names with a sticky note and ask, 'Do you see a change?'

02At a glance

Intervention
not applicable
Design
other
Sample size
36
Finding
not reported

03Original abstract

Masked visual analysis (MVA) was developed to complement traditional visual analysis (TVA) and control for Type I error rates. Researchers have empirically tested MVA with generated data and simulated decisions. Our purpose was to evaluate the performance of MVA with real data and human raters. We asked visual analysts who had published at least one single‐case research article (n = 36) to evaluate nine multiple‐baseline‐design‐across‐participants graphs. Graphs representing different target behaviors were displayed in masked and unmasked presentations. We evaluated the reliability and validity of MVA and TVA. Agreement within each method was similar to that reported in previous studies on visual analysis (MVA ICC = 0.625; TVA ICC = 0.579). Between the two methods, at least 75% of raters' decisions corresponded for six of nine graphs. We discuss the implications of incorporating MVA and future research on analytic methods for single‐case experimental designs.

Journal of Applied Behavior Analysis, 2026 · doi:10.1002/jaba.70043