Evaluation of components of residential treatment by Medicaid ICF-MR surveys: a validity assessment.
Medicaid clipboard surveys trigger staff reactivity and hide quality gaps—use brief, unannounced behavioral checks instead.
01Research in Context
What this study did
Goldstein et al. (1991) checked if Medicaid ICF-MR surveys tell the truth. They compared survey scores for certified and non-certified units in large state homes for adults with intellectual disability.
The team looked at whether higher survey scores matched better client care. They also watched if staff acted differently when surveyors were in the building.
What they found
Surveys gave the same scores to both good and bad units. Certified units scored no better than non-certified ones.
Staff cleaned up only when surveyors arrived. This reactivity masked real quality gaps and made the survey useless.
How this fits with other research
Guercio et al. (2025) show a fix. They paid staff with small preferred items for turning in ≥80% of daily data sheets. Fidelity jumped in every house, proving you can get clean data if you motivate staff the right way.
Waller et al. (2010) saw the same reactivity risk. Month-end integrity checks clustered like schedule-induced scallops, warning that any announced observation can warp staff behavior.
Emerson et al. (2013) widen the lens. They caution that self-report and proxy data for people with ID often carry hidden validity flaws, backing the 1991 warning to doubt clipboard metrics.
Why it matters
Stop trusting clipboard surveys. If you need true quality data, use brief, unannounced behavioral samples or reward staff for accurate daily data like Guercio et al. did. Spread integrity checks across the month and keep them short to avoid the scallop trap. When you write reports, flag any measure that could be faked—your funding source and clients deserve real numbers, not theater.
Want CEUs on This Topic?
The ABA Clubhouse has 60+ free CEUs — live every Wednesday. Ethics, supervision & clinical topics.
Join Free →Pick one key client response, define a 5-minute spot-check, and do three unannounced counts this week.
02At a glance
03Original abstract
We evaluated the proficiency of the federal Medicaid program's survey process for evaluating intermediate care facilities for the mentally retarded. In Study 1, an observational analysis of active treatment during leisure times in living units suggested that these surveys did not discriminate between certified and noncertified units. In Study 2, a reactivity analysis of a survey indicated that direct-care staff performed differently during the survey by increasing interactions with clients and decreasing nonwork behavior. Similarly, results of Study 3 showed increases in client access to leisure materials during a survey. In Study 4, questionnaire results indicated considerable variability among service providers' opinions on the consistency, accuracy, and objectivity with which survey teams determine agency standard compliance. Results are discussed regarding effects of the questionable proficiency of survey processes and the potential utility of behavioral assessment methodologies to improve such processes.
Journal of applied behavior analysis, 1991 · doi:10.1901/jaba.1991.24-293