By Matt Harrington, BCBA · Behaviorist Book Club · April 2026 · 12 min read
Treatment fidelity — the degree to which an intervention is implemented as designed — is not simply a quality assurance checkbox. It is a foundational variable that mediates the relationship between a behavior analytic program and its outcomes. When procedures are implemented with low fidelity, the data collected during those sessions cannot be interpreted as reflecting the actual effect of the intervention. Decisions based on that data are, at best, imprecise. At worst, they lead to program modifications that address an implementation problem as if it were a learner problem.
Patricia Glick's work presenting aggregated industry fidelity data from 29 organizations provides something the field has rarely had access to: a population-level view of how ABA is actually being delivered across diverse settings. Single-organization fidelity studies tell us what one organization achieves under its specific conditions. Multi-organization data reveals whether those achievements are typical, exceptional, or below the industry standard — and where the gaps between prescribed and delivered treatment are most prevalent.
For service delivery at the organizational level, fidelity data serves multiple functions. It is a measure of staff competence, a measure of supervisory effectiveness, a measure of training quality, and ultimately a measure of the organization's commitment to its clients. Organizations that collect fidelity data systematically and use it to drive continuous improvement are operationalizing their quality standards in measurable terms. Organizations that rely on supervisor impressions or client progress as proxies for fidelity are missing a critical layer of process monitoring.
The BACB Ethics Code addresses this directly. Code 2.19 (maintenance of client records and data) and Code 4.07 (ongoing data collection) together establish that behavior analysts must collect and use data to drive clinical decisions. Fidelity data is part of that obligation — it is process data that contextualizes outcome data. A client who is not making expected progress on a skill acquisition program may be experiencing a learning problem, a program design problem, or an implementation problem. Without fidelity data, these three possibilities cannot be distinguished.
At the industry level, understanding where fidelity commonly breaks down — across settings, provider types, client profiles, and program types — creates a shared knowledge base that individual organizations can use to benchmark their own practices and identify improvement targets.
The concept of treatment fidelity entered the behavior analytic literature primarily through the research methodology literature, where it was recognized that published studies needed to demonstrate that experimental conditions were implemented as described. A study claiming to demonstrate the effectiveness of functional communication training is uninterpretable if implementation varied widely from the prescribed protocol — the effect attributed to FCT may reflect a different intervention. This methodological concern translated over time into a clinical concern: if research findings only hold when procedures are implemented accurately, clinical practice should measure and maintain that accuracy.
Fidelity measurement tools in ABA typically involve direct observation against a structured checklist of critical procedural components. For discrete trial teaching, fidelity checks might include whether the discriminative stimulus was presented clearly, whether the correct prompt level was used, whether reinforcement delivery was contingent and immediate, and whether inter-trial intervals were appropriate. For naturalistic intervention, fidelity checks might assess whether the therapist was following the learner's lead, embedding teaching opportunities appropriately, and using the prescribed response strategies.
The challenge Glick's research addresses is that most organizations collect fidelity data, but far fewer aggregate it systematically to identify patterns across programs, staff, and time. Fidelity data collected in isolated binders or disconnected spreadsheets cannot reveal the organizational or industry-level trends that would allow for systemic improvement. Cross-organizational analysis requires data infrastructure, common measurement frameworks, and willingness to share data across organizational boundaries — all of which are relatively rare in a field that has historically operated in fragmented, competitive service markets.
The findings from 29 organizations provide a benchmark reference that allows individual providers to understand their fidelity rates in context. An organization achieving 75% average fidelity on direct instruction programs may not know whether that figure is typical or problematic without an industry reference point. If the multi-organization data shows that well-performing organizations achieve 90%+ on comparable measures, that gap becomes actionable information.
Provider-level versus client-level analysis is a key distinction in this research. Provider-level fidelity averages can mask important variation: a provider whose average across clients is 80% may achieve 95% with some clients and 60% with others, with the variation driven by client complexity, program type, or setting factors. Client-level analysis reveals where individual learners are receiving substandard implementation — a more clinically relevant finding than an organizational average.
The direct clinical implication of fidelity data is straightforward: clients receiving high-fidelity treatment make more progress than clients receiving low-fidelity treatment on the same programs. This has been demonstrated repeatedly in the skill acquisition literature for specific procedures including discrete trial training, video modeling, pivotal response training, and functional communication training. The effect sizes associated with implementation accuracy are large enough to be clinically meaningful — not just statistically significant.
For individual BCBAs, this means that fidelity monitoring of the staff implementing their programs is not an optional supervisory task. It is a required component of ethical service delivery. A BCBA who designs a behavior intervention plan but never directly verifies that it is being implemented as written has incomplete information about what is actually happening in their client's sessions. Code 4.07's requirement for ongoing data collection applies to process data as well as outcome data — a program that is showing no progress should prompt fidelity review before program modification.
Fidelity data also has implications for how program modifications are made. The behavior analytic decision-making process typically involves reviewing outcome data and adjusting the program when progress is insufficient. If fidelity data shows that implementation accuracy is low, modifying the program in response to poor outcome data may be the wrong decision — the appropriate response is improving implementation, not changing the program. Organizations that do not collect fidelity data routinely may be making program modifications that treat a staff performance problem as a learner problem, which wastes resources and delays effective intervention.
At the supervision level, fidelity data provides objective performance feedback that is more useful than subjective supervisor impressions. Telling a therapist that their sessions are going well, or that they need to be more consistent, conveys little actionable information. Presenting a therapist with their fidelity scores by session, by program, and compared to their previous scores provides specific data they can act on. This is the difference between feedback and coaching — fidelity data enables the latter.
For organizations managing multiple programs across diverse client profiles, aggregated fidelity data by program type reveals which procedures are implemented most and least consistently across the organization. Procedures with systematically low fidelity may indicate that the training for those procedures is insufficient, that the fidelity measure itself is not sensitive enough, or that the procedure is too complex to implement reliably without additional support structures.
The ABA Clubhouse has 60+ on-demand CEUs including ethics, supervision, and clinical topics like this one. Plus a new live CEU every Wednesday.
The BACB Ethics Code creates clear obligations around fidelity that many practitioners do not explicitly recognize as fidelity requirements. Code 2.09 requires the use of least restrictive, effective procedures. Determining whether a procedure is working requires knowing that it is being implemented correctly — a procedure that appears ineffective because of poor fidelity may be prematurely abandoned, denying a client an intervention that would have worked with accurate implementation. Fidelity data is the mechanism that distinguishes these scenarios.
Code 2.12 addresses the obligation to advocate for resources necessary to implement behavior analytic services effectively. If an organization's fidelity data shows that therapist-to-supervisor ratios, caseload volumes, or training resources are insufficient to maintain acceptable fidelity levels, BCBAs have an ethical obligation to advocate for changes to those conditions. Documenting fidelity problems provides the evidence base for that advocacy — it transforms a qualitative concern about service quality into a measurable organizational performance problem.
Informed consent has a fidelity dimension that is often overlooked. When families are presented with a behavior intervention plan and consent to its implementation, they are consenting to that specific procedure implemented in the specified manner. If implementation consistently departs significantly from the plan, families are not receiving what they consented to. This is an informed consent issue, not merely a quality issue. Code 2.03 requires ongoing informed consent — which arguably includes informing families when the intervention they consented to is not being delivered with the fidelity that would make it effective.
Data integrity is another ethics concern. Code 2.15 requires honest and accurate data collection. Fidelity data that is collected through self-report, or by observers who are not blind to expected performance levels, is subject to social desirability bias that inflates apparent fidelity. Organizations committed to ethical data practices must design fidelity measurement systems with appropriate safeguards — trained, calibrated observers; inter-observer agreement checks; observation schedules that include unannounced visits.
Sharing industry fidelity data across organizations, as this research does, also raises data ethics questions. De-identification, appropriate consent from participating organizations, and responsible reporting of findings are essential for this kind of research to be conducted ethically and for its results to be trusted.
Designing a fidelity measurement system requires decisions about what to measure, how to measure it, who does the measuring, and how often. The target behaviors for fidelity checklists should map directly to the critical procedural components of the intervention — not every step, but the steps whose omission or modification would most significantly affect the learner's response to the procedure. Overly long fidelity checklists are burdensome to complete and less likely to be implemented consistently; focused checklists with 8-15 items per procedure tend to generate more usable data.
Observation frequency affects both the reliability of fidelity estimates and the feasibility of the measurement system. Too infrequent — monthly or less — provides a snapshot that may not represent typical implementation. More frequent brief observations — several times per month — capture more of the natural variation in performance and allow earlier detection of drift. Many ABA organizations use a tiered observation schedule: more frequent observation for new staff and new programs, less frequent for established staff with demonstrated fidelity histories.
Inter-observer agreement (IOA) for fidelity data is often neglected but is essential for data interpretability. If two observers watching the same session disagree significantly on whether components were implemented correctly, the fidelity measure has a reliability problem that undermines its validity as a quality indicator. IOA checks should be conducted regularly and calibration sessions scheduled when agreement drops below acceptable thresholds.
Strategic planning based on fidelity data should follow a decision tree: first, identify programs and staff with fidelity below the organizational target; second, analyze whether the problem is at the staff skill level (training problem), the supervisory feedback level (coaching problem), or the systems level (resource or infrastructure problem); third, implement the corresponding intervention. Using fidelity data only to identify problems without a structured decision process for addressing them produces organizational anxiety without improvement.
For benchmarking against industry data from studies like Glick's, organizations should ensure that their fidelity measurement tools are sufficiently comparable to the measurement approaches used in the research. Organizational-level benchmarking is most valid when the measurement methodology is aligned.
The strategic plan for enhancing fidelity in your organization starts with a current-state assessment. Pull your existing fidelity data, or conduct a brief audit of fidelity measurement practices if data is not currently being collected systematically. Identify which programs are being monitored, how often, by whom, and what the data shows. That baseline is the starting point for improvement.
If your organization is not currently collecting fidelity data systematically, the first priority is not choosing the right tool — it is building the measurement habit. A simple, consistently applied fidelity checklist for your two or three highest-priority procedures, used by trained observers on a defined schedule, produces more useful information than a sophisticated tool used sporadically.
Connecting fidelity data to outcome data is the analytical step that transforms fidelity monitoring from a quality assurance activity into a clinical decision-making tool. When you see a client whose skill acquisition is stalling, routinely check fidelity data alongside program data. When fidelity is high and progress is slow, you have a program design problem. When fidelity is low, you have an implementation problem. The distinction determines the intervention, and it cannot be made without both data streams.
Sharing aggregated, de-identified fidelity data within your organization — across teams, programs, and sites — allows the internal benchmarking that drives continuous improvement. Staff and supervisors who see their fidelity data in context make more informed decisions about where to invest development effort.
Ready to go deeper? This course covers this topic in detail with structured learning objectives and CEU credit.
Industry fidelity data as an indicator of quality service delivery — Patricia Glick · 1 BACB Supervision CEUs · $0
Take This Course →All behavior-analytic intervention is individualized. The information on this page is for educational purposes and does not constitute clinical advice. Treatment decisions should be informed by the best available published research, individualized assessment, and obtained with the informed consent of the client or their legal guardian. Behavior analysts are responsible for practicing within the boundaries of their competence and adhering to the BACB Ethics Code for Behavior Analysts.