Assessment & Research

Artificial intelligence for the measurement of vocal stereotypy

MM et al. (2020) · 2020
★ The Verdict

An AI audio model can track vocal stereotypy as accurately as a human observer for most kids with autism, cutting observation load.

✓ Read this if BCBAs who measure vocal stereotypy in clinic or school settings
✗ Skip if Practitioners working only with motor stereotypy or adults without vocal reps

01Research in Context

01

What this study did

MHeald et al. (2020) trained an AI to listen for vocal stereotypy in kids with autism. They recorded eight children during regular sessions. A neural network learned to flag the repetitive sounds.

Human observers also scored the same tapes. The team then compared the AI counts to the people counts, session by session.

02

What they found

The AI matched the humans on at least eight out of every ten sessions for six of the eight kids. In other words, the computer agreed with trained staff 80% of the time or better for most participants.

The result shows a cheap laptop can do the tallying instead of a tired clinician.

03

How this fits with other research

Maharaj et al. (2020) did the same trick with a Kinect camera. Their system tracked motor stereotypy with 92% accuracy. Together, the two studies prove AI can watch or listen and still hit the 90% club.

Gilchrist et al. (2018) and Keintz et al. (2011) used small accelerometers on wrists or torsos. They caught hand-flapping and body-rocking at 80-90% accuracy. MM et al. extend this line by adding the audio channel, so now both movement and sound can be scored without human timers.

Shawler et al. (2020) looked at how to reduce vocal stereotypy once you measure it. Their RIRD treatment paper pairs nicely with MM’s measurement paper—you can let the AI count while you run the intervention.

04

Why it matters

You no longer need a second staff member glued to a stopwatch. Set a tablet on the desk, hit record, and the AI logs vocal stereotypy for you. Use the saved staff time to teach instead of tally. Start small: record one client for one session and check the AI count against your own. If the numbers line up, let the model carry the load next week.

Free CEUs

Want CEUs on This Topic?

The ABA Clubhouse has 60+ free CEUs — live every Wednesday. Ethics, supervision & clinical topics.

Join Free →
→ Action — try this Monday

Record five minutes of a client’s session, run the free AI script, and compare its count to yours—if it matches, you just gained an extra pair of ears.

02At a glance

Intervention
not applicable
Design
single case other
Sample size
8
Population
autism spectrum disorder
Finding
positive
Magnitude
medium

03Original abstract

Both researchers and practitioners often rely on direct observation to measure and monitor behavior. When these behaviors are too complex or numerous to be measured in vivo, relying on direct observation using human observers increases the amount of resources required to conduct research and to monitor the effects of interventions in practice. To address this issue, we conducted a proof of concept examining whether artificial intelligence could measure vocal stereotypy in individuals with autism. More specifically, we used an artificial neural network with over 1,500 minutes of audio data from 8 different individuals to train and test models to measure vocal stereotypy. Our results showed that the artificial neural network performed adequately (i.e., session-by-session correlation near or above .80 with a human observer) in measuring engagement in vocal stereotypy for 6 of 8 participants. Additional research is needed to further improve the generalizability of the approach.

, 2020 · doi:10.1002/jeab.636