This guide draws in part from “Exploring the Use and Implications of Artificial Intelligence in the Practice of Behavior Analysis” by Sara Gershfeld, BCBA (BehaviorLive), and extends it with peer-reviewed research from our library of 27,900+ ABA research articles. Citations, clinical framing, and cross-links below are synthesized by Behaviorist Book Club.
View the original presentation →Artificial intelligence is no longer a speculative technology for applied behavior analysis. AI-powered tools are already being marketed to ABA providers for clinical documentation, data analysis, treatment planning, and progress monitoring. This course examines both the potential benefits and the significant challenges that AI integration presents for behavior analytic practice, with particular attention to how these technologies affect clinical decision-making, treatment efficiency, and the practitioner-client relationship.
The clinical significance of this topic is multifaceted. On the opportunity side, AI systems can process large volumes of session data far more quickly than human analysts, potentially identifying patterns in behavior that would take clinicians much longer to detect. AI can automate administrative tasks such as session note generation, insurance documentation, and scheduling, freeing clinician time for direct client interaction. Some AI systems offer real-time feedback to therapists during sessions, which could enhance treatment fidelity and accelerate skill development for new practitioners.
On the risk side, the introduction of AI into clinical practice raises fundamental questions about the nature of clinical judgment. The BACB Ethics Code (2022) is built on the premise that a qualified human professional is making informed decisions about client care. When AI systems influence or automate portions of that decision-making process, the locus of clinical responsibility becomes unclear. Who is accountable when an AI system recommends a treatment modification that produces a negative outcome? How does the behavior analyst evaluate whether the AI's recommendation is appropriate for this specific client in this specific context?
Core Principle 2.01 (Providing Effective Treatment) requires behavior analysts to rely on scientific evidence and professional judgment. AI-generated recommendations may or may not align with the evidence base, and clinicians may lack the technical expertise to evaluate the algorithms underlying AI outputs. Core Principle 2.13 (Accuracy in Billing and Reporting) raises questions about how AI-assisted documentation is reviewed and verified. If an AI system generates session notes that the clinician signs without careful review, the clinician bears ethical responsibility for any inaccuracies.
The pace of AI development far outstrips the pace of professional regulation. The BACB Ethics Code does not explicitly address AI, and state licensing boards have been slow to issue guidance. This creates a regulatory vacuum in which practitioners must rely on general ethical principles to navigate specific technological challenges. This course provides a framework for doing so.
The intersection of artificial intelligence and behavior analysis exists within a broader healthcare trend. AI and machine learning technologies are being adopted across medical, psychological, and educational fields with varying degrees of evidence support and regulatory oversight. In some domains, such as radiology and dermatology, AI systems have demonstrated performance comparable to human specialists for specific tasks. In others, AI implementations have produced biased, inaccurate, or clinically harmful outputs.
Within ABA, AI applications are emerging across several domains. Data collection and analysis tools use machine learning to identify patterns in behavioral data, predict skill acquisition trajectories, and flag potential regression. Natural language processing systems generate clinical documentation from session recordings or therapist inputs. Computer vision systems are being explored for automated behavior coding from video recordings. Recommendation engines suggest treatment modifications based on client data and population-level outcomes.
Sara Gershfeld brings expertise in the practical realities of ABA practice to this discussion, grounding the conversation in how these technologies affect frontline clinicians and the clients they serve. This practical perspective is essential because much of the discourse around AI in healthcare is driven by technology developers rather than practitioners.
The historical context is important. Behavior analysis has always valued measurement and data-driven decision-making. In many ways, the field is better positioned than most to integrate AI tools because it already generates the structured behavioral data that AI systems require. However, the field's commitment to individualized assessment and treatment creates tension with AI systems that derive recommendations from population-level patterns. What works for most clients may not work for this client, and the behavior analyst's ability to make that distinction is precisely what AI cannot replicate.
The regulatory landscape is evolving but underdeveloped. As of this course, neither the BACB nor most state licensing boards have issued specific guidance on AI use in behavior analytic practice. The Health Insurance Portability and Accountability Act (HIPAA) applies to AI systems that process protected health information, but compliance requirements for AI-specific risks such as algorithmic bias or model drift are not well-established. Practitioners are operating in an environment where the technology is ahead of the guidance.
There is also a market context. AI tools are being actively marketed to ABA providers, often with claims about improved efficiency, reduced burnout, and better outcomes. These claims may or may not be supported by rigorous evidence, and practitioners need critical evaluation skills to distinguish genuinely useful tools from overhyped products.
The integration of AI into clinical practice has implications for every phase of service delivery, from assessment through discharge. Understanding these implications requires distinguishing between AI as a tool that augments clinical decision-making and AI as a system that replaces it.
In assessment, AI systems can analyze historical data to identify patterns that inform diagnostic impressions and treatment planning. For example, machine learning algorithms applied to behavioral data may detect subtle correlations between environmental variables and challenging behavior that a clinician might miss. However, these correlations are statistical associations, not functional relationships. The behavior analyst must still conduct a functional assessment to determine whether the identified patterns reflect true maintaining variables or spurious correlations. Over-reliance on AI-generated assessments could lead to treatments that target the wrong variables.
In treatment planning, AI recommendation engines may suggest goals, interventions, or modifications based on what has worked for similar clients in the system's database. The clinical implication is that these recommendations reflect population-level trends and may not account for the unique cultural, familial, and contextual factors that shape each client's needs. The behavior analyst's responsibility is to evaluate each recommendation against their knowledge of the individual client and to override AI suggestions when their clinical judgment indicates a different course of action.
In data analysis, AI tools can process session data more rapidly than manual methods, generating graphs, identifying trends, and flagging potential concerns in near-real-time. This has clear benefits for clinical responsiveness. However, it also creates a risk of data overwhelm. When a system generates dozens of automated insights per week, the clinician must be able to distinguish clinically meaningful patterns from noise. This requires strong data analysis skills that may atrophy if practitioners become dependent on AI interpretation.
In documentation, AI-generated session notes and treatment reports can reduce administrative burden significantly. The clinical risk is that clinicians may sign off on AI-generated documentation without sufficient review, leading to inaccuracies in the clinical record. Because the clinical record is the primary source of information for treatment continuity, insurance authorization, and legal proceedings, errors in AI-generated documentation can have serious downstream consequences.
In staff training and supervision, AI tools that provide real-time feedback during sessions may accelerate therapist skill development. However, they may also undermine the supervisory relationship if technicians begin to rely on automated feedback rather than engaging in reflective discussions with their supervisors. The development of clinical judgment requires grappling with uncertainty, considering multiple perspectives, and learning from mistakes, processes that automated feedback cannot fully replicate.
The ABA Clubhouse has 60+ on-demand CEUs including ethics, supervision, and clinical topics like this one. Plus a new live CEU every Wednesday.
The ethical considerations surrounding AI in behavior analysis are extensive and largely uncharted by formal ethical guidance. Practitioners must extrapolate from the general principles and standards of the BACB Ethics Code (2022) to navigate specific AI-related challenges.
Data privacy is a primary concern. AI systems that analyze clinical data typically require that data to be transmitted to and stored on external servers, often controlled by third-party technology companies. Core Principle 2.06 (Maintaining Confidentiality) requires behavior analysts to take reasonable precautions to protect client information. This obligation extends to understanding how AI vendors handle client data, including where it is stored, who has access to it, whether it is used to train the vendor's models, and what happens to the data if the vendor is acquired or goes out of business. Many practitioners lack the technical expertise to evaluate these factors, creating a gap between ethical obligations and practical capacity.
Algorithmic bias is another significant concern. AI systems learn from historical data, and if that data reflects existing biases in service delivery, the AI will perpetuate those biases in its recommendations. For example, if the training data disproportionately represents clients from certain demographic backgrounds, the system's recommendations may be less accurate or appropriate for clients from underrepresented populations. Core Principle 1.07 (Cultural Responsiveness and Diversity) requires behavior analysts to be responsive to diversity, which in the AI context means evaluating whether the tools they use produce equitable recommendations across populations.
Informed consent presents unique challenges. Core Principle 2.04 requires that clients and stakeholders be informed about the nature of services. When AI tools are used in assessment, treatment planning, or data analysis, clients have a right to know that their data is being processed by automated systems, how those systems work in general terms, and how AI-generated recommendations influence their care. Most informed consent documents do not currently address these issues.
Competence boundaries are also relevant. Core Principle 1.04 (Practicing within a Defined Role) and 1.05 (Practicing within Scope of Competence) require behavior analysts to operate within the boundaries of their training and expertise. Most behavior analysts have limited training in computer science, machine learning, or data engineering. Using AI tools without understanding their capabilities and limitations may constitute practicing outside one's competence. This creates an obligation for ongoing professional development in AI literacy.
The clinician's responsibility for AI-influenced decisions is perhaps the most fundamental ethical issue. Regardless of what an AI system recommends, the behavior analyst who implements that recommendation bears professional and ethical responsibility for the outcome. This principle must be maintained even as AI tools become more sophisticated and their recommendations more persuasive. The Ethics Code does not permit delegation of clinical responsibility to an algorithm.
When considering whether and how to integrate AI tools into your practice, a structured decision-making framework can help you navigate the complexity. This framework should address the tool's clinical utility, its ethical implications, your competence to use it, and its impact on the practitioner-client relationship.
The first step is to evaluate the evidence base for the specific AI tool. Ask the vendor for peer-reviewed research supporting the tool's effectiveness and accuracy. Be skeptical of claims supported only by internal company data or testimonials. Determine whether the tool has been validated with populations similar to your clients. If the evidence is insufficient or unavailable, weigh whether the potential benefits justify the uncertainty.
The second step is to conduct a data privacy assessment. Identify what client data the tool collects, where it is transmitted, how it is stored, and who has access. Determine whether the tool's data practices comply with HIPAA and any applicable state privacy laws. Read the vendor's privacy policy and terms of service carefully, paying attention to clauses about data ownership, model training, and data sharing with third parties.
The third step is to assess your own competence. Do you understand how the tool generates its outputs? Can you evaluate whether those outputs are clinically appropriate for a given client? If not, seek training or consultation before implementing the tool. Consider whether your organization can provide ongoing technical support and whether your supervisees and staff have the training needed to use the tool appropriately.
The fourth step is to update your informed consent process. Before using AI tools with clients, update your consent documents and conversations to include information about what AI tools you use, what data they process, how their outputs influence clinical decisions, and the client's right to opt out. This is not a bureaucratic exercise but a genuine effort to ensure that clients and families understand and agree to the use of technology in their care.
The fifth step is to establish oversight procedures. AI tools should augment clinical decision-making, not replace it. Establish protocols for reviewing AI-generated outputs before they influence clinical decisions. This includes reviewing AI-generated documentation for accuracy, evaluating AI treatment recommendations against your clinical judgment and knowledge of the individual client, and monitoring for signs of algorithmic bias or drift over time.
Ongoing monitoring is essential. AI systems can change as they are updated by vendors or as they learn from new data. What was accurate and appropriate six months ago may not be today. Establish a schedule for reviewing the performance and appropriateness of any AI tools you use, and be prepared to discontinue tools that no longer meet your clinical and ethical standards.
AI is coming to behavior analysis whether individual practitioners embrace it or not. The question is not whether you will encounter AI tools in your practice but how prepared you will be to evaluate, implement, and oversee them responsibly.
Start by building your AI literacy. You do not need to become a computer scientist, but you do need to understand the basics of how machine learning works, what training data is, how algorithmic bias can arise, and what the limitations of AI-generated outputs are. This knowledge is becoming as essential to competent practice as understanding research methodology.
Approach AI vendor claims with healthy skepticism. Ask for evidence. Ask who funded the research. Ask what populations were included in validation studies. Ask about data privacy practices. The behavior analytic emphasis on evidence-based practice applies to the tools you use, not just the interventions you deliver.
Update your informed consent procedures now, before AI tools become standard. Clients and families deserve to know when AI is involved in their care and how it is being used. This is both an ethical obligation under the BACB Ethics Code (2022) and a trust-building practice that strengthens the therapeutic relationship.
Maintain your clinical judgment as the primary driver of decision-making. AI outputs are inputs to your thinking, not substitutes for it. When an AI system recommends something that does not align with your knowledge of the client, trust your clinical expertise. Document your reasoning and move on.
Finally, advocate for the development of professional guidance on AI use in behavior analysis. Engage with your state licensing board, your professional association, and the BACB to push for clear, practical guidelines that help practitioners navigate this rapidly evolving landscape.
Ready to go deeper? This course covers this topic in detail with structured learning objectives and CEU credit.
Exploring the Use and Implications of Artificial Intelligence in the Practice of Behavior Analysis — Sara Gershfeld · 1 BACB Ethics CEUs · $20
Take This Course →We extended this guide with research from our library — dig into the peer-reviewed studies behind the topic, in plain-English summaries written for BCBAs.
279 research articles with practitioner takeaways
258 research articles with practitioner takeaways
252 research articles with practitioner takeaways
You earn CEUs from a dozen different places. Upload any certificate — from here, your employer, conferences, wherever — and always know exactly where you stand. Learning, Ethics, Supervision, all handled.
No credit card required. Cancel anytime.
All behavior-analytic intervention is individualized. The information on this page is for educational purposes and does not constitute clinical advice. Treatment decisions should be informed by the best available published research, individualized assessment, and obtained with the informed consent of the client or their legal guardian. Behavior analysts are responsible for practicing within the boundaries of their competence and adhering to the BACB Ethics Code for Behavior Analysts.