Guidelines From the American Society of Pain and Neuroscience for Using Artificial Intelligence in Interventional Spine and Nerve Treatment

Introduction History of Artificial Intelligence

The phrase Artificial Intelligence (AI) originated in 1956 at the Dartmouth Summer Research Project Artificial Intelligence conference, where pioneers Allen Newell, Cliff Shaw, and Herbert Simon introduced Logic Theorist, which is thought to be the first AI program.1 Although this early concept showed a promising foundation, AI research faced severe setbacks in the late 1970s into the early 1990s, known as the “AI Winter”, due to perceived technological limitations, reduced funding, and insufficient databases.2

The earliest applications of AI in medicine were in the 1970’s. They included CASNET, which could apply disease information to patients and provide management advice, and MYCIN, which used patient information to provide potential bacterial pathogens and recommended treatments.2 Starting in the 1990’s, the advancement of computing power and machine learning techniques reignited the evolution of AI, which accelerated in subsequent decades with the advent of techniques to integrate unstructured data. With the evolution of these systems, AI has flourished into more mainstream applications such as virtual assistants and healthcare systems. Today, we are faced with balancing the use of ongoing research to integrate these systems into daily practice while ensuring responsible development and utilization of these models.

Technical Fundamentals

AI relies on data and math models. At the core is an extensive data repository that trains and develops algorithms that predict categories or outcomes based on a range of inputs.

Machine Learning

Machine learning strategies broadly fall into the categories of supervised and unsupervised. Data is divided into training, validation, and test data.3 Training data is used to develop the model, validation data is used to fine tune and validate the data, and test data is unseen data used to evaluate the model. Supervised learning uses labeled data to make predictions based on inputs.4 Labeled data are data points that are labeled with the category that applies. An example would be data on patients with sacroiliitis who have had SI joint fusion labeled as patients with successful outcomes and those without. Machine learning requires large volumes of data, and a shortage of labeled data in the medical field is a significant obstacle.

The simplest mathematical form of supervised learning is linear regression where a line of best fit is used to predict the values of the dependent variable (Figure 1).5 It performs well for data with a linear relationship. Logistic regression is used for binary data such as 0 or 1 or True or False. The value of the dependent variable represents the probability of the outcome (Figure 2). However, both methodologies tend to overfit and may not apply to more complex relationships.

Figure 1 Linear Regression model. A “best fit” line predicts the outcomes of the dependent variables.

Figure 2 Logistic Regression Model, used for binary data where the dependent variable represents the probability of the outcome.

Support Vector Machines (SVM) rely on plotting data points and identifying a hyperplane, essentially a line, when in two dimensions, dividing the data between two categories (Figure 3).6 The optimal hyperplane maximizes the probability of correct classification by maximizing the distance or margin between the hyperplane and any point in the training set.4 When data is not linearly separable, a kernel function can plot a complex three-dimensional representation such that a hyperplane can exist to separate the categories.

Figure 3 Support Vector Machines. These algorithms are used for classification and regression. They use a line (hyperplane) to delineate between categories.

Decision tree algorithms use yes/no questions to sort data into categories.3 They are prone to overfitting and small changes in training data can significantly change predictions. Random forest algorithms rely on many decision trees built on training data samples. Each tree provides an outcome which functions like a “vote” for that output. The output with the most “votes” becomes the model’s output. These algorithms can handle nonlinear data and implicitly perform feature selection. However, they are computationally expensive.

Unsupervised learning is a method where algorithms learn from unlabeled data.6 One purpose is clustering. Imagine that there are multiple phenotypes of patients with spinal stenosis. k-Means clustering could identify multiple subgroups within the patient population.4 Centroids are plotted, and the sum of squared distances between the data points and centroids is calculated. The number of centroids and locations are varied through progressive iterations until the optimal centroids are identified. Thereafter, any new point will belong to the group associated with the centroid closest to it. A better understanding of the types of patient groups may help with later data exercises to determine if certain groups are better suited to interventions.

Neural Networks

Neural networks are mainly supervised learning.6 Each input is a node represented by a value and then passed through a connection to a node in the next layer of the neural network after being multiplied by a weight associated with the neuron connecting the nodes and a bias that is added or subtracted from each node. Neural networks consist of an input and an output layer (Figure 4). Deep neural networks contain additional hidden layers between the input and output layers. Training data is used to adjust the weights and biases throughout the neural network with each iteration of training data passing through the network. Gradient descent and backpropagation are utilized to massively decrease the time and computational power required to train a neural network.

Figure 4 Neural Networks. Each node creates an output that is associated with the next node, which creates another output to the next layer. This occurs at every layer of the network.

Reinforcement learning is employed to determine optimal behavior where multiple courses of action are possible in sequential steps.4 Reward and value functions evaluate the success of various strategies and solve very complex problems, particularly where the outcomes of actions are not always predictable Reinforcement learning is very successful in medicine and has been applied to protein folding problems.

Transfer learning uses a model trained on one task to rapidly develop a model for a second task with less data.6 Deep learning models frequently employ transfer learning. An example would be using a model that understands spoken words and then using transfer learning to train this model to understand medical terminology. Generative Adversarial Networks (GAN) use a pair of neural networks, a generator and a discriminator to train for tasks.7 The generator attempts to fool the discriminator by outputting fake data. The discriminator, trained on the desired outcome, evaluates and grades the generator’s output to determine if the output data is real or fake. Through repetitive iterations, it is possible the generator will fool the discriminator with fake data.

Large Language Models (LLM) utilize massive volumes of data to learn the meaning of words in various contexts and generate meaningful text.8 ChatGPT, Gemini, and Grok are well-known examples. OpenGPT is an excellent example of an implementation that can be used for transfer learning to customize for unique applications such as within the medical field.

Creating an AI model starts with preparing data (Figure 5). This encompasses tasks ranging from cleaning up data and addressing missing data to reducing dimensionality. Subsequent tasks consist of training, packing, and validating the model. Deploying models is more complicated than simply uploading the final model. Deployment at scale may require pruning of neural networks for size reduction and optimization of models to reduce the computation power required. The financial and computational costs of AI models can multiply quickly at scale. Finally, models must be monitored to improve performance over time and prevent drift.

Figure 5 AI Model Workflow. Begins with preparing data, followed by training, packaging, and validating the model before deployment. After deploying the model, it must be monitored for accuracy.

General Applications in Medicine

Artificial Intelligence (AI) is revolutionizing medicine with diverse applications. Machine learning algorithms assist in early disease detection, such as identifying cancerous lesions in medical imaging.9 Natural Language Processing (NLP) enables efficient electronic health record analysis, reimbursement optimization, automated note generation, and patient-doctor communication.10 AI-powered robotic surgery enhances precision and minimizes invasiveness.11 Drug discovery has benefited from AI’s ability to identify promising compounds and predict their efficacy.12 Personalized medicine leverages AI to tailor treatments based on patient data.13 AI’s predictive capabilities aid in disease outbreak forecasting and resource allocation. AI-driven digital likeness technology may enable clinicians to scale their patient education and engagement efforts, inside and outside the clinic setting. AI can summarize or even develop scholarly research and aid in peer review.14 These applications demonstrate AI’s transformative potential for healthcare delivery and patient outcomes.

Regulation and Limited Guidance

No federal legislation explicitly limits the use of artificial intelligence (AI) and machine learning (ML), but several documents providing guidance have been presented for the general use of AI/ML. President Biden issued an Executive Order on “Safe, Secure, and Trustworthy Artificial Intelligence” which focused on creating safe and responsible uses of AI.15 It required AI systems developers to share safety test results with the government if results showed the AI could pose a risk to national security.15 This was immediately rescinded by President Trump with “Removing Barriers to American Leadership in Artificial Intelligence” which revoked AI policies thought to hinder American innovation in AI.16 It also calls for the creation of an Artificial Intelligence Action Plan to “sustain and enhance America’s global AI dominance in order to promote human flourishing, economic competitiveness, and national security”.16 This highlights the dynamic nature of legislation surrounding AI based on different administrative goals. The “Blueprint for an AI Bill of Rights: Making Automated Systems Work for the American People” was published by the White House Office of Science and Technology Policy in October 2022.17 The AI Bill of Rights comprises five principles (Table 1).

Table 1 AI Bill of Rights

Regulation Through the Food and Drug Administration (FDA)

In 2021, the FDA issued the “Artificial Intelligence/Machine Learning (AI/ML)-Based Software as a Medical Device (SaMD) Action Plan”.18 (Box 1). Consistent with the action plan, the FDA later issued documents on how adjustments and changes to AI/ML are made safely and effectively (Table 2). The FDA also provides a list of AI/ML-enabled medical devices marketed in the United States.

Box 1 AI/ML SaMD Action Plan

Table 2 FDA Documents on How AI/ML Changes Are Made

In March 2024, the FDA published the “Artificial Intelligence and Medical Products: How CBER, CDER, CDRH, and OCP are Working Together”, outlining their coordinated approach to AI. Despite these documents, there remains an overall paucity of guidance and regulations regarding developing and deploying these technologies, specifically within the medical field.

Regulation Through The US Department of Health and Human Services (HHS)

HHS released a Strategic Plan for the Use of Artificial Intelligence in Health, Human Services, and Public Health outlining their vision to be a leader in innovating and adopting responsible AI to advance American health and well-being.19 The framework and roadmap ensures HHS fulfills its obligation to the nation and pioneers the responsible use of AI.

The Strategic Plan is organized into seven domains, including five primary domains and two functional areas:

Primary Domains:

Medical Research and Discovery Medical Product Development Healthcare Delivery Human Services Delivery Public Health

Functional Areas:

Cybersecurity and Critical Infrastructure Protection Internal Operations

Each domain chapter follows a consistent structure, discussing stakeholders, opportunities, trends, potential use cases and risks, and an action plan.

HHS supports a coordinated public-private approach to improving quality, safety, efficiency, accessibility, equitability, and outcomes in health and human services through innovative, safe, and responsible development and use of AI.19 The plan focuses on four key goals:

Catalyzing health AI innovation and adoption Promoting trustworthy AI development and ethical use Democratizing AI technologies and resources Cultivating AI-empowered workforces and organization cultures.

The Strategic Plan acknowledges both AI’s potential benefits and risks in health and human services. It emphasizes the importance of responsible adoption, addressing bias, privacy, and security issues. The plan also highlights the need for collaboration among various stakeholders, including government agencies, academic institutions, industry partners, and patients, to ensure AI’s successful and ethical implementation in healthcare and related fields. ASPN endorses this strategic vision, and this Neuron Project provides the first societal framework to support its adoption.

Methods

The American Society of Pain and Neuroscience (ASPN) established a multidisciplinary panel to develop expert consensus on the application of artificial intelligence (AI) within pain management. This panel consisted of leading pain medicine physicians, anesthesiologists, physical medicine and rehabilitation physicians, and neurosurgeons, augmented by promising residents and fellows with significant interest in the field. Critical focus areas were identified and subsequently refined by the lead and senior authors, alongside a group of section editors. Authors were then organized into teams to conduct comprehensive literature reviews and develop their respective sections. To ensure impartiality, all authors disclosed financial conflicts of interest and were required to recuse themselves from any discussions or content related to their competing interests. In instances where a conflicted author possessed unique expertise in a specific area, a non-conflicted author assumed the ultimate editorial oversight for the submitted material.

The main objective of the panel was to create a systematic guideline process to outline the current literature on artificial intelligence (AI) in relation to the practice of pain medicine. This paper provides a primer on AI to the pain practitioner to help them understand AI to utilize in their practice. A literature search was conducted using PubMed to evaluate the current literature regarding AI relevant to pain practitioners. Search terms included key words such as “pain”, “pain medicine”, “pain management”, “analgesia”, “opioid”, and other relevant terms cross-referenced with “artificial intelligence”, “machine learning”, “deep learning”, and “neural network”. Two reviewers screened results to remove articles that were not relevant to our review. For example, articles discussing AI that were not primarily focused on pain management and vice versa. A deep dive into the literature to evaluate the quality of presented evidence was not performed, as it was deemed outside the scope of this paper. Broad categories of articles were created using the available data and reference papers, broad categories of articles were created. These categories give a general idea of the goals and use cases for articles discussing AI in pain management.

With these broad categories in mind, senior reviewers met and discussed important aspects of AI that practitioners must understand before implementing AI in clinical practice. These aspects include understanding the potential benefits of AI, the potential harms of AI, and how to ensure the benefits of AI outweigh the risks. With these aspects clearly defined, authors were divided into groups, and another literature search was conducted to review the current literature regarding the benefits and harms of AI. Authors utilized their literature searches to complete each section within the framework.

The guideline development process commenced with section editors and lead authors collaboratively outlining key principles for each focus area. An initial comprehensive draft of guidelines was then consolidated, with lead and senior authors integrating analogous themes and eliminating redundant content. This list of candidate guidelines underwent an iterative, multi-cycle refinement process by the entire author group, leading to a streamlined, heavily edited collection of guidelines that had reached a convergence of shared expert opinion. The final, votable guideline list was subsequently distributed for unanimous approval from all contributing authors.

Results

Over the past 20 years, the number of publications regarding AI has increased exponentially. Searching “artificial intelligence” in PubMed before 2000 produces just over 3000 results. From 2021 until the time of writing, over 71,000 articles have been published discussing artificial intelligence (Table 3). We found over 12,000 articles pertinent to pain management and artificial intelligence using our search terms. Most articles about chronic pain and artificial intelligence were not clinical studies but are reviews offering guidance to clinicians.20–22 The remaining articles cover broad topics in pain management, utilizing AI to improve the field. Major topic categories include aiding in the diagnosis of pain, monitoring, management, outcomes, and enhancing research capabilities.

Table 3 Number of PubMed Search Results for “Artificial Intelligence” (as of Oct 30, 2024)

We found no significant articles offering broad understanding and guidelines applying artificial intelligence in pain medicine. This paper fills those gaps and gives the clinician the tools to understand and use AI in their daily practice.

The AI healthcare market has experienced explosive growth in recent years, with projections indicating continued rapid expansion. According to a recent report, the global AI in Healthcare market was valued at USD 20.9 billion in 2024 and is estimated to reach USD 148.4 billion by 2029. This represents a compound annual growth rate (CAGR) of over 38% during the forecast period.

Key factors driving this remarkable growth include:

Increasing adoption of AI technologies in medical imaging, diagnostics, and drug discovery Rising demand for personalized medicine and improved patient outcomes Growing investments in AI healthcare startups and research initiatives Advancements in machine learning algorithms and computing power Expanding applications in telemedicine and remote patient monitoring.

While estimates vary slightly between sources, the consensus points to a market poised for significant expansion. One report projects AI in the healthcare market to surpass USD 187.95 billion by 2030, highlighting the immense potential for AI to revolutionize healthcare delivery and management in the coming years.23–25

Conclusions/Recommendations/Consensus Guidelines for AI in Pain Medicine Benefits of Artificial Intelligence - Enhancing Healthcare AI in Diagnosis and Treatment Planning Treatment Algorithms and Patient Selection

Common healthcare professional challenges include accurate identification of patients benefiting from a specific therapy and the time required to review a patient’s history. Specific time constraints or unfamiliarity with interpreting different image modalities can lead to an over reliance on radiology reports, which may not discuss crucial pathologies relevant to chronic disease states that have established FDA-approved treatments. Additionally, administrative and payor-related regulations can lead to unsuccessful authorization of beneficial therapies.

Artificial Intelligence (AI) and Machine Learning (ML) technologies evolve and “learn” through data feeding, a process that allows models to update and refine their algorithms, leading to more accurate and effective learning over time. When used as a support tool, these technologies help implement personalized treatment algorithms based on different types of variables within those data flows. Examples include genetic variants, patient history, relevant physical exams, imaging analysis, ICD codes, labs, and medication regimens. This feature not only has the potential to engender more appropriate patient management but also liberate clinicians’ time, allowing for enhanced patient care.26

Given the significant promise of these technologies, several companies have focused on building machine learning into medical decision making. They are creating tools to support clinicians, as well as algorithms designed to function independently of them.

Medical Imaging Interpretation

One of the most promising applications of ML within healthcare is in the field of radiology. For instance, ML has been successfully used to detect and predict lung malignancy in 8000 cases, outperforming non-thoracic radiologists.27 Artificial Intelligence (AI) algorithms, especially those based on deep learning, can analyze complex medical images, identify patterns invisible to the human eye, and provide accurate diagnoses. This significantly enhances the efficiency and accuracy of medical imaging interpretation, reducing the workload of clinicians interpreting various imaging modalities, while potentially improving patient outcomes. ML can substantially enhance the radiology workflow when combined with Electronic Health Records (EHRs) and clinical decision-making processes. It could potentially automate dose estimation, improve examination quality control, and expedite the turnaround for radiographic images and interpretation.28 This would boost efficiency and enhance precision in identifying pertinent positive findings.

With the advent of digitized images and the availability of massive imaging datasets, coupled with advancements in computer vision, machine learning performance is anticipated to surpass human accuracy.29 These algorithms are not subject to fatigue, and model accuracy continues to improve as these algorithms are fed with uninterrupted data.

These algorithms could also be particularly beneficial in medical imaging interpretation for preoperative and intraoperative planning. For example, in a sacroiliac joint fusion, massive datasets can analyze multiple anatomic variables simultaneously to determine the precise positioning for the implantation of allograft or titanium screws. Based on treatment success variables, postoperative imaging model datasets can also enhance implant positioning.

Improving Treatment Equity

Not all treatment plans are created equally, with disparities in treatment quality that can be attributed to various factors such as clinician inexperience, patient selection challenges, issues with workflow algorithms, or implicit bias. Implicit biases are subconscious prejudices that can be difficult to recognize and standardize against. They contribute to disparities in health care access, quality of care, and overall health outcomes.30

Given that ML models are fundamentally data-driven, model design may incorporate these biases and should thus be scrutinized for their presence. To counter such biases, data for algorithms should be balanced and acquired from a wide range of diverse populations to reduce health care disparities and improve patient outcomes.31

Prediction of Outcomes

ML applications have demonstrated effectiveness in disease prediction utilizing diverse data types, medical imaging, and clinical outcome prediction.32–34 These sophisticated technologies can process and analyze vast amounts of data, identifying patterns and making remarkable predictions. This predictive power can significantly influence medicine and clinical decision-making processes.35 By offering insights into the success rates of different treatment types, ML applications may assist doctors in determining the most effective treatment plan. For patients, especially those faced with various treatment options, these evidence-informed data points are critical to provide a reasonable estimate of success, enabling patients to make more informed and confident decisions about their health.

Staying Up to Date with Current Literature

In the rapidly evolving landscape of medicine, the immense complexity of emerging therapies demands a solution to comprehend better the volume of evidence generated daily. The surge in data-driven treatment solutions necessitates integrating incorporating AI more than ever. AI plays a pivotal role in various aspects of healthcare, particularly in diagnosis and treatment recommendations. These are primarily based on a comprehensive review of the latest literature. AI tools that process and interpret vast amounts of data will offer invaluable service in summarizing published studies by learning from patterns and adapting based on feedback.

ML excels in managing many predictors, often surpassing the number of observations and amalgamating them in nonlinear and highly interactive ways. This capability enables us to utilize new data types, the volume or complexity of which would have previously rendered them unmanageable for analysis.36 The learning component of these systems is critical as they can be trained to continually enhance their diagnostic precision and clinical decision-making recommendations aligned with the most recent data.

Treatment and Treatment Optimization Electrical Delivery Individualized

For spinal cord stimulation, clinicians currently collaborate with device specialists to adjust parameters such as waveform, frequency, amplitude, and others to tailor electrical flows in stimulation. This practice inherently leads to variability between the programming of these parameters without a mechanism to reliably determine the optimal stimulation settings for every component. The use of AI/ML in spinal cord stimulation presents a significant opportunity to refine practice among the increasing numbers of SCS implants annually.37 One of the most common medical applications of AI/ML is neural network “learning” (ie, being trained) by processing examples containing “inputs” and “results”.38 When considering its application in SCS, inputs may include stimulation parameters, patient data (age, sex, BMI, comorbidities, spinal anatomy), diagnosis, and pain pattern. Results may include longitudinal outcomes such as pain scores and patient-reported outcome measures (PROMs). Ultimately, such use of AI/ML would predict and inform optimal settings for the delivery of stimulation for each individual patient, minimizing variability in current practice.

Medication Delivery Individualized

Similar limitations exist in medication delivery for pain management. The trialing and titrating of medications varies from clinician to clinician, and may lead to delayed progress, excess burden on healthcare systems, side effects, and inferior outcomes. Using of AI/ML to process existing data (inputs and results) may predict optimal medication(s) and dose for each specific patient and calculate the risk of side effects, misuse, and inefficacy. High-quality, high-volume data would allow for incremental learning and refinement of such computational algorithms to drive improved patient outcomes.38

Virtual Reality

Virtual, augmented, and mixed reality have several applications in interventional and surgical specialties.39 One application is their use in training. Some interventional pain management fellowships may not provide adequate exposure to neuromodulation and advanced innovative procedures in the field.40 These technologies allow trainees to interact with a virtual or mixed environment to practice operative techniques and sharpen interventional skills, improving accuracy, faster operative time, and reducing complication rates.39 Another application of these technologies is their intraoperative use. AR/MR projects graphics onto real-world surfaces for enhanced intraoperative guidance. With the use of patient preoperative imaging, these technologies have been shown to assist in the precise placement of percutaneous sacroiliac joint screws, reduce fluoroscopy and operative time, as well as facilitate accurate spinal instrumentation.39,41,42 As the footprint of sacroiliac joint fusion/stabilization, interspinous process devices, and spinal arthrodesis expands in interventional pain management, the application of AR/MR intraoperatively with VR presents an opportunity to improve techniques, outcomes, and patient safety.

Identification of Complications

The use of AI-driven models to predict and identify complications also has growing use-cases and substantial promise. For example, an AI model trained on the data of 3034 patients and 35 unique input variables to predict surgical site infection after posterior spinal fusion had a positive predictive value of 92.6%.43 Applying AI to predict complications after endovascular aneurysm repair led to a model with 100% sensitivity for post-operative complications.43 Such use can readily apply to interventional pain management, where patient data and inputs may be utilized to create similar predictive models for post-operative complications such as infection, lead migration, wound dehiscence, and other significant outcomes. Higher-risk patients could then be considered for closer postoperative surveillance, thus avoiding potential delays in care. Furthermore, advancements in patient-wearable devices may help provide AI models with key inputs to identify complications such as lead migration, lead fracture, or implantable pulse generator (IPG) failure in real-time. Patient-wearable data may provide input variables for AI models targeting activity, steps, and heart rate.44 AI/ML could then synthesize the complex relationships in changes and variations between such inputs and others over time to create a model for real-time complication detection.

Injections

Integrating AI/ML with fluoroscopy also presents several opportunities for optimizing injection-based interventions. The potential reduction in radiation exposure has been highlighted in several studies.45–47 For example, in a study comparing AI-equipped fluoroscopy with conventional fluoroscopy in endoscopy, AI-enabled fluoroscopy significantly decreased radiation exposure by using ultrafast collimation targeting the area of interest.46 In oncology and pathology, deep learning analysis of the critical features of disease in imaging has accelerated more accurate and earlier detection of cancers.48,49 Incorporating such deep learning with AI-enabled fluoroscopy systems in interventional pain management has many potential applications. This includes real-time detection and confirmation of contrast spread in the intended treatment area during an injection and high-sensitivity detection of unwanted spread (ie intravascular, intrathecal). In this manner, utilization of AI can optimize outcomes and safety with fluoroscopy-based treatments.

Safety AI and Opioid Safety

For opioid management, Large Language Models (LLMs) may efficiently synthesize and retrieve information from an expansive corpus of medical literature, clinical guidelines, and patient records, contributing to better decision-making. With more optimized opioid dosing and therapeutic outcomes, opioid-associated risks including abuse and overdose can be better mitigated to deliver safer, more effective care.50

Patient education is critical to safe opioid use. AI chatbots educating patients with materials tailored to their specific needs may reduce opioid use post-operatively and increase patient satisfaction. Applications of AI chatbots have already augmented the patient experience while maximizing provider time.51 Several studies have examined LLM application in post operative pain and demonstrated positive outcomes in opioid-related adverse events.51 These applications are useful for practitioners without specialized training who manage post operative pain. Additionally, these models are being used in opioid abuse treatment and show promise in predicting retention with treatment.52

For opioids, wearable technologies providing biometric data interpreted by AI can now use cardiac and respiratory sensors to assess opioid use, overuse, and overdose, including designs with a naloxone auto-injector.53 While still early in development, LLM applications with real-time biometric responses to pain medications and/or interventions have the potential to fundamentally alter opioid safety in both acute and chronic pain settings.

Patient Engagement Workflow Efficiency

Pain management clinicians frequently grapple with time-consuming administrative tasks, including documentation in EMRs and ensuring accurate coding for billing purposes. These challenges impede workflow efficiency and have financial implications for healthcare organizations. Integrating AI technologies offers promising solutions to address these challenges, enabling clinicians to devote more time to patient care while optimizing practice management processes.54

AI integration with EMRs holds immense potential in streamlining workflow processes for pain management clinicians. Natural Language Processing (NLP) algorithms can automate the documentation process by extracting relevant information from clinical notes, reducing the time spent on manual data entry. Furthermore, AI-powered decision support systems can analyze patient data in real-time, providing clinicians with actionable insights and personalized treatment recommendations. By automating repetitive tasks and offering intelligent decision support, AI integration enhances workflow efficiency, allowing clinicians to focus on delivering high-quality patient care.55

Improving Coding Accuracy with AI-Assisted CPT Coding

Accurate coding is crucial for pain management practices to ensure proper reimbursement and compliance with regulatory requirements. However, the complexity of CPT coding often leads to errors and inconsistencies, resulting in revenue loss and compliance risks. AI-assisted CPT coding tools leverage machine learning algorithms to analyze clinical documentation and recommend appropriate codes based on the procedures performed and the documentation provided. These tools improve coding accuracy and facilitate compliance with billing regulations, reducing the risk of audits and revenue loss.

Research

Large language models (LLM) have been suggested to aid in clinical research.56 LLM-based chatbots can assist with preliminary research design, including providing an overview of the evidence, defining a research question and hypothesis, and calculating the needed sample size for a powered study. During the clinical trial, LLM chatbots may assist with patient-trial matching, coding free-text narratives, and detecting cognizant consent. Following data collection completion, LLMs may even execute statistical analysis of data and technical writing of the manuscript. However, it is essential to note, that despite the LLM providing answers or solutions to these research tasks, this does not necessarily mean the answer is correct or appropriately conducted. Thus, at present, chatbots do not replace scientific training and experience but rather should be regarded, at most, as a starting point that requires validation.57

Optimize Healthcare Administration

Pain management clinicians face unique challenges in healthcare administration, including navigating prior authorizations, optimizing scheduling, and promoting their services effectively in competitive markets. Inefficient administrative processes can result in delays, increased costs, and reduced patient satisfaction. Consequently, there is a growing need to explore innovative strategies for optimizing healthcare administration in pain management settings.

Optimization of Prior Authorizations

Prior authorizations are often required for medications, procedures, and therapies in pain management practices, adding administrative complexity and delaying patient care. Implementing electronic prior authorization (ePA) solutions and utilizing AI-powered tools can streamline the authorization process, reduce paperwork, and expedite approvals. By integrating ePA systems with electronic health records (EHRs) and leveraging AI algorithms to predict authorization requirements, pain management clinicians can minimize administrative burdens and enhance patient access to timely treatments.58

Streamlining Scheduling Processes

Efficient scheduling is crucial for optimizing patient flow, minimizing wait times, and maximizing resource utilization in pain management clinics. Utilizing scheduling software with customizable features, automated appointment reminders, and real-time patient tracking capabilities can improve scheduling efficiency and patient satisfaction. Moreover, implementing telemedicine options for follow-up appointments and leveraging patient portals for self-scheduling can further streamline the scheduling process, offering convenience and flexibility to patients while reducing administrative overhead for clinicians.59

Effective Marketing Strategies

Artificial Intelligence holds immense potential to transform marketing strategy development in pain management practices, enabling personalized patient engagement, targeted outreach, and service optimization. By harnessing AI-driven insights, practices can enhance patient satisfaction, loyalty, and outcomes while optimizing resource allocation and revenue generation. However, addressing data privacy, interpretability, and ethical use challenges is crucial for successfully integrating of AI into pain management marketing strategies. Nonetheless, the opportunities presented by AI offer a pathway to more effective and patient-centric pain management practices in the digital age.

New Drug Product Development

Regarding basic and translational science, LLM may accelerate drug discovery by predicting the structural target, location of binding, pharmacodynamic properties, pharmacokinetic properties and dosages associated with therapeutic benefit versus toxicity.60 Once the medication is developed, clinicians and patients can use LLM to identify drug-drug interactions60 and more rapidly advance through the phases of clinical trials. Several biomedical companies are already employing LLM in product development.60

Potential Harms of AI

Artificial intelligence has amassed significant attention and excitement due to its potential applications in enhancing the medical community and optimizing healthcare for both clinicians and patients. However, we must consider potential pitfalls, harms, and limitations in leveraging AI. By acknowledging these challenges, we can ensure that AI is used safely, meaningfully, and equitably while preserving the integrity and humanity of healthcare delivery. This section examines potential concerns regarding the use of AI in medicine:

Hallucinations/Errors

One inherent risk associated with AI is the possibility of generating hallucinations or errors in data processing. AI algorithms, particularly machine learning models, rely heavily on data input for training and decision-making. The quality of AI output is contingent upon the quality and comprehensiveness of the data used to train the model. Machine learning algorithms are therefore only as proficient as the data they are exposed to. This raises the concern for bias, reporting inaccuracies, and data inadequacies. One example is IBM’s Watson supercomputer, aka Watson, which uses AI algorithms to generate treatment recommendations for cancer patients. However, the data used to create Watson was based on a few hypothetical cases with oncologist input. Recommendations made by Watson were sometimes erroneous and had the potential to harm a patient if executed.61

AI tends to fabricate or extrapolate data to fill in gaps in existing knowledge. This is known as data hallucination. Data hallucination can lead to erroneous conclusions, particularly if not reviewed for accuracy. AI has also been known to fabricate references that do not exist. For instance, Alkaissi & McFarlane utilized ChatGPT to generate a paragraph on homocysteine-induced osteoporosis.62 While the AI generated an articulate and seemingly well-rounded discussion on the topic, a closer review revealed inaccuracies in some conclusions. Further, when asked to provide references to support the AI-generated content, ChatGPT supplied citations that did not exist, with PubMed IDs corresponding to unrelated papers.62

Such hallucinations and errors in AI can propagate misinformation and false assumptions, ultimately compromising patient care and the integrity of scientific knowledge.

Over-Reliance on AI

Another significant concern surrounding AI is over-reliance on automated systems which can have far-reaching consequences. In medicine, diagnoses are often complex and not black and white, usually requiring deductive reasoning and experience to delineate. For instance, two patients with the same type of pain may describe it very differently as pain perception of pain is colored by many components including cultural and emotional dimensions. It is unclear how AI will account for these more subjective elements.

Over-reliance on AI can also lead to diminished patient-clinician relationships as reduced human interaction can cause patients to feel alienated. Wu et al, conducted over 3000 patient surveys and found that the perception of clinician empathy and trust in the clinician’s benevolence significantly influenced the patient’s evaluation of the patient-clinician relationship.63 Trust, communication, and empathy are cornerstones in medicine and cannot be diminished in delivering compassionate, patient-centered care.

Studies regarding applications of artificial intelligence in medicine are difficult to regulate due to the complexity of the data. Additionally, with the increasing popularity of generative AI to write scientific narratives, properly evaluating presented data is more critical than ever. A study using deep learning to improve the diagnosis of spinal foraminal stenosis has been retracted for discrepancies in scope, data, and inappropriate citations.64

Another concern is inequitable resource allocation. Vulnerable populations with low technological literacy or socioeconomic status may face barriers to accessing AI-driven healthcare services. These innovative services may disproportionately be utilized amongst privileged groups, widening health inequity.65

Ethical Considerations

Determining accountability becomes paramount when considering failures within AI and ML systems, posing a significant ethical quandary. China and Hong Kong’s prohibition of AI in medical decision-making underscores the urgency of addressing accountability. The absence of universal guidelines in healthcare AI usage fosters moral and ethical debates, prompting ongoing research by organizations like the FDA and NHS to establish standards. However, regulatory approval and legal frameworks for AI and ML actions remain challenging for authorities and courts.66

Model Bias

Bias within the medical realm can be analyzed from three perspectives: data-driven, algorithmic, and human biases. In healthcare AI, such biases can have a negative impact by perpetuating societal prejudices ingrained within datasets. Consequently, patient groups historically underrepresented, such as gender and ethnic minorities, risk misdiagnosis, exacerbating existing disparities.

Future endeavors should prioritize establishing AI standards in healthcare that prioritize transparency and data sharing, all while safeguarding patient privacy. Adopting open science methodologies can promote fairness in healthcare AI. These encompass: (1) involving participants in AI algorithm development and embracing participatory science; (2) promoting responsible data sharing and inclusive data standards to ensure interoperability; and (3) facilitating code sharing, including algorithms capable of synthesizing underrepresented data to mitigate bias.67

Data Privacy

One inherent concern with AI and healthcare is the large amount of sensitive data required for AI models to analyze patient data and potentially predict treatment outcomes effectively. The inherent complexity of medical reasoning and the underlying heterogeneity of datasets necessitates extensive data sets to build reliable models. The challenge with AI in healthcare is related to acquiring a large dataset, and more importantly, to the sensitive nature of this data. AI requires access to sensitive electronic medical records, imaging scans, laboratory results, and other personal data to generate recommendations. Laws to protect patient confidentiality such as HIPAA, have been in effect for nearly three decades. However, these laws were enacted long before the advent of AI in modern-day healthcare. Conflict over ownership of health data has also emerged with AI, where patients are thought to deserve to own and control their own data, but often other members of the healthcare and AI systems make claims for control of the information.68

The concern is ensuring integrity and patient confidentiality are maintained using AI models. This was also a concern during the transition from paper charts to electronic health records (EHR). Although EHRs have undoubtedly revolutionized healthcare in many ways, the electronic nature of these records makes them susceptible to significant breaches. A 2020 US Department of Health and Human Services report states that approximately 578 healthcare institutions experienced data breaches, impacting over 41 million individuals.69 One study suggested that up to half of the US population may have had their medical records compromised.70 The concern is even more amplified when appreciating that this same vulnerable data may need to be moved into other storage environments and aggregated across multiple institutions to reach the necessary scale. Larger datasets represent an enormous trophy for bad actors.

Another potential harm of AI in healthcare is the lack of transparency concerning acquisition, distribution, usage and interpretation of data. The “black box” phenomenon is a well-documented criticism of AI models. Artificial intelligence’s inability to explain or justify its complex conclusions can add unwanted ambiguity to an already sensitive matter: a patient’s health. As clinicians, we need to know not just the “what” but also the “why” and “how”, which can be difficult to ascertain at times with this type of technology. Additionally, data that is too much for a non-AI model might be too little for a specific AI algorithm. Therefore, building a comprehensible explanation of an AI model is often impossible. There is growing work in “E xplainable AI”, which is a technique to help humans understand and trust AI systems s. Finally, although there remains a push towards an increase in regulatory oversight on AI data distribution, the rapid advancement of AI has thus far outpaced most regulatory bodies, adding another layer of unclarity of how sensitive patient data is being distributed. There remains a need for large-scale global initiatives that help standardize the laws governing patient data distribution in healthcare-related AI systems.

Limitations to Data in Healthcare

Although healthcare data is rapidly growing and becoming more accessible, it still lags the massive scale of big data available outside the healthcare industry. Accuracy is essential for big data analytics. There have been some limitations due to issues with data collection. For example, personal records may contain typing errors or abbreviations. Medical data input placed by a medical assistant may contain errors or misspellings that alter meaning. There may also be an underrepresentation of groups in health care data due to issues such as literacy and language barriers, resulting in data sets that may not be representative.

There are many economic challenges in collecting big data as well. Smaller healthcare enterprises are poorly equipped to modernize data collection systems. Most healthcare organizations are highly fragmented, preventing universal data collection. Security and privacy issues pose another. Finally, bureaucratic limitations, heavy regulation, and security concerns limit data aggregation and availability.

Legal Challenges to AI Use Copyright Infringement

With the rapid evolution of AI in medicine, we must understand the current copyright laws in place to protect both creators and their creations. In the United States, copyright law grants the creator ownership of the work. Copyright protection supports the original works of authorship fixed in a tangible medium of expression.71 One of the exceptions to copyright protection is fair use, which allows a creation to be used without notifying or getting permission from the author for criticism, commentary, news reporting, teaching, scholarship, or research purposes.72 The Copyright Act of 1976 gives ownership to the natural person or persons who created the work unless created within the scope of work or person(s) or as “work made for hire”. This is defined as “a work specially ordered or commissioned for use as a contribution to a collective work, as part of a motion picture or other audiovisual work, as a translation, as a supplementary work, as a compilation, as an instructional text, as a test, as answer material for a test, or as an atlas”.73 It is essential to know that there is a significant exception to this general rule found in Section 201 (b) of the Copyright Act. This states “in the case of a work made for hire, the employer or other person for whom the work was prepared is considered the author for purposes of this title, and, unless the parties have expressly agreed otherwise in a written instrument signed by them, owns all the rights comprised in the copyright”. To summarize, the employer owns the copyright of a creation that is AI generated content made at work or under “work made for hire”. This leads to the question of who owns AI-generated content when an AI system autonomously generates it without direct human input. Some jurisdictions may consider the AI system developer as the owner, while others may consider the AI system itself or the user of the AI system the owner.74 There have been several cases in the United States where AI systems were used as joint authors of a work under US law. One case in the Southern District of New York declared that only a natural person can be the inventor.74 We must also consider copyright protection of original work. This depends on various factors, including the degree of human involvement in the creation process. In March of 2023, a notice of proposed rulemaking was published that allowed works containing AI-generated material to be eligible for registration under specific circumstances.74 The United States Copyright Office has declared that a human author must produce copyright material. Therefore, any material created by AI is, by definition, uncopyrightable because it does not require human creativity.75

Because AI models require the consumption of large quantities of data, copyright infringement issues come to the forefront when training large systems with inputs that the AI developer does not directly own. This has been well publicized in the case of Meta allegedly downloading over 80 terabytes of copyrighted materials such as books and textbooks, to train their LLaMa AI.76 Employees even had extensive discussion questioning whether this was legal.

As medical implementations of AI move forward, these types of concerns regarding the underlying datasets that power models will quickly become relevant. There are certainly arguments to be made that transparency as to the entirety of inputted works and their provenance will instill greater confidence in utilizing AI models within healthcare. Furthermore, they will allow outsiders to assess flaws with the underlying knowledge set or risk for bias.

Lack of Oversight Oversight in AI Use in Pain Medicine

Artificial intelligence (AI) has been rapidly integrated into various fields of medicine, including pain management, promising significant advancements in diagnostics, treatment personalization, and patient monitoring. However, the pace of AI development and implementation has often outstripped the ability of regulatory frameworks to keep up, resulting in limited oversight. This gap poses potential risks, including biases in AI algorithms, privacy concerns, and the unintended consequences of AI-driven decisions on patient care.

Current oversight mechanisms for AI in pain medicine primarily revolve around existing medical device regulations and healthcare data privacy laws. In the United States, the Food and Drug Administration (FDA) has started to develop frameworks specifically for AI and machine learning (ML) in healthcare. The FDA’s Digital Health Innovation Action Plan outlines the agency’s approach to fostering innovation while ensuring patient safety.77 This includes a proposed regulatory framework for AI/ML-based Software as a Medical Device (SaMD), emphasizing the need for continuous learning and adaptation of AI systems while maintaining rigorous safety and effectiveness standards.77 However, the oversight remains fragmented and often reactive rather than proactive. The approval process for AI tools is not always transparent, and there is a lack of standardized methods for evaluating AI’s performance in real-world clinical settings.77 Moreover, many AI applications in pain medicine are not strictly regulated as medical devices, such as decision support tools or predictive analytics platforms, which can fall through the regulatory cracks.78

Future Regulatory Issues in AI Use in Pain Medicine

Integrating AI into pain medicine holds immense promise but also presents significant challenges. Limited oversight, rapid expansion, and potential unintended consequences highlight the urgent need for robust regulatory frameworks that can keep pace with technological advancements. By addressing critical issues proactively, the medical community can harness the benefits of AI while safeguarding patient safety and equity. One major challenge is the establishment of comprehensive guidelines for AI development, validation, and deployme

Comments (0)

No login
gif