OpenAlex · Aktualisierung stündlich · Letzte Aktualisierung: 20.03.2026, 07:11

Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.

Realistically Integrating Machine Learning Into Clinical Practice: A Road Map of Opportunities, Challenges, and a Potential Future

2020·64 Zitationen·Anesthesia & AnalgesiaOpen Access
Volltext beim Verlag öffnen

64

Zitationen

4

Autoren

2020

Jahr

Abstract

Recently, interest in machine learning has rapidly grown within the health care community. This popularity has been due in part to the advances in perception tasks such as speech processing and image analysis, which has led to the successful implementation of machine learning across a variety of industries: from the ancient game of Go to Netflix movie suggestions.1 The abundance of electronic health record data coupled with the low costs of computation power and data storage makes machine learning and medicine well matched. Although health care has been traditionally slow to adopt new technologies, the time is approaching when machine learning delivers on promises to improve the current and future practice of perioperative medicine. Deployment and operationalization of a machine-learning model require synthesizing knowledge in data processing and model development with a knowledge of medicine and clinical workflows. In this article, we hope to elucidate what machine learning is and why it will transform clinical care, discuss what it takes to implement machine learning in clinical care, address current limitations and drawbacks, and ultimately examine what the future of machine learning in health care may hold. A PRIMER ON MACHINE LEARNING Machine learning is a subset of artificial intelligence in which a computer iteratively learns from data without explicit rule-based programming. Machine-learning models find patterns within data and apply these patterns to new data to make predictions. Because of the computer’s ability to rapidly process large amounts of electronic data, these models are advantageous when using large datasets. Where humans can become overwhelmed with increasingly large and complex data inputs, machine-learning models often thrive. The recent rapid explosion in computational processing power and increasing availability of large data enable capabilities beyond traditional rule-based modeling, including improved analytical capacities and accessibility to unique, potentially hidden insights. As these methods become more widely adopted in the medical community, all parties will benefit. Improvements include increased care efficiency and more accurate hospital reimbursements and increase care quality, such as improved disease classification, predicting complications, and ultimately better outcomes, Machine-learning models have been created to predict an increasing number of clinical outcomes, such as diagnoses and mortality, with applications including C. difficile infection in the inpatient hospital setting,2 identifying molecular markers for cancer treatments,3 and postoperative surgical outcomes.4 Examples of machine learning include a cardiologist using an automated interpretation of an ECG and a radiologist using an automated detection of a lung nodule in a chest x-ray. In both of these examples, a machine-learning model approximates a trained physician’s diagnosis with high accuracy. To date, models are successfully constructed to answer a single clinical question by using appropriately labeled representative data. The tools to develop these models (using languages such as R and Python) are largely free of charge and openly accessible. However, the data required to train these models and the expertise in deploying them remain hurdles to widespread adoption. We are currently in a state of discovery, with most efforts centered around model development. Unfortunately, a substantial number of models fail to progress to clinical deployment. The promise of machine-learning applications and their appropriate implementation remain separated, with progression slowed by inexperience, a lack of direction, and the need for multiple complimentary skill sets that may not be present in a single institution or team. The focus on model development can overshadow important evaluations of the appropriateness of the question being addressed, model scrutiny, and technical or political roadblocks in operationalizing results. To be successful, it is important to consider costs, impact, outcomes (including moral and ethical), and path to deployment before projects begin. MODEL IMPLEMENTATION The key step in developing and operationalizing a machine-learning model is picking a good problem to target. Ideally, these problems have some important characteristics: they are regarded as being of high clinical or monetary value, they represent areas of medicine or operations where there is consensus by providers on the appropriate management, the necessary data are already available, and the application of computing power has the promise to reduce cognitive burden for the provider. Few problems will have all of these properties. For instance, a system has been described that predicts intraoperative hypotension, based on high-fidelity analysis of arterial waveforms.5 Thus, using data already available, providers can potentially intervene on this significant pathophysiological state before it develops. However, the appropriate therapy, whether it is administration of volume, delivery of vasopressors, or adjustment of the anesthetic, would still need to be determined. The next step is to create a team of individuals capable of creating the desired solution.6 While abilities and roles may overlap, the team will need clinical expertise, data experts, and design help. Clinical expertise requires both a clinical subject matter expert and a clinical champion to assist with implementation. To address data needs, the team will need an individual capable of extracting appropriate datasets from the electronic health record, someone to perform data validation, and an expert in machine-learning models. Some individuals may play more than one role. For instance, a clinical subject matter expert who is facile with data extraction and has time for extensive data validation could get much of the work done. With an addressable problem and a team in place, the next step is to develop a good solution. Machine-learning tools must obey the 5 rights of decision support: providing the right information, to the right person, in the right format, through the right channel, at the right time in the workflow.6 Conceptualizing the intervention requires understanding the implications of model performance. For example, one of the authors recently evaluated a model in operation that predicted the need for a postacute care bed at the time of discharge, with great performance: area under the curve (AUC) at 0.87, and sensitivity and specificity of roughly 80%. With approximately 94% of patients discharged to home, however, that translates to a positive predictive value of 24% and a negative predictive value of 75%.7 When used to drive a list of patients expected to have postacute care needs, it performs poorly: 3 of 4 patients flagged are going home, and 1 in 4 going to postacute care are missed, resulting in wasted time. Predicting patients going home instead yields a much better result: a positive predictive value of 98%, which case managers can use to reliably identify patients who do not need to be evaluated, thus saving significant time. Finally, piloting the solution with a limited group of users is critical. This can be done in a production environment by restricting the number of users or areas to which the solution is available or by releasing the solution into a separate test or support environment for evaluation. Use of a production environment may yield better feedback but may not always be feasible or may be much more resource intensive. After a successful pilot, the solution is ready for deployment. It is important to embed an evaluation strategy into the rollout to determine whether the solution was effective. It is often helpful to share the results of the implementation with the end users and to periodically reevaluate deployed solutions, which have performance that may degrade over time as patient characteristics or practice patterns change with time, making iterative changes as necessary.8 In addition, solutions may develop malfunctions over time, and solutions no longer working appropriately should either be fixed or decommissioned.9 CHALLENGES TO DEVELOPMENT AND IMPLEMENTATION There are a number of hurdles that are still present in machine-learning development and implementation in anesthesiology. Many of these challenges are similar to those that have plagued outcomes and database researchers for some time. The first set of drawbacks relate to model development. While the expansion in data availability is beneficial for creating machine-learning models, the variety of data sources, as well as the trials of data wrangling, can be formidable. Incompatible data formats from different sources, unusual data structures, and generally messy data can occupy a significant amount of time in the development process.10 Similarly, if models are created using a single institution’s data, there still remains the issue of external validation from multiple sources.11 Fortunately, using data from large registries may be helpful in addressing these concerns. First, large multicenter data warehouses, such as those built by the Anesthesia Quality Institute, Multicenter Perioperative Outcomes Group, and National Surgical Quality Improvement Program, typically attempt to unify disparate data sources and formats into a common format.12 Second, they contain data from multiple institutions. This allows for testing of generalizability by running a given model on data from different institutions. While there are still data acquisition and quality difficulties with large multicenter warehouses, depending on how they are structured and assuming institutions submit quality data, the standardization of data elements is a step in the right direction for health care–specific data management. To impact patient care, models must not only be developed but also deployed. Given a lack of interoperability standards, for a model to work, it must be able to interface with the data within different EHRs. Each EHR contains varying data structures, creating a significant challenge to model deployment. It is similar to having a recipe in one’s language and the ingredients in another and not knowing whether the right ingredients are in the pantry. Unfortunately, competition between EHR companies, between hospitals, and even between silos within an institution limit the drive for interoperability. Recently, the Office of the National Coordinator for Health Information Technology updated their meaningful use criteria to include interoperability standards; however, this is a more recent change.13 Hopefully, this will force the various stakeholders (including those in competition) to restructure systems to improve interoperability. Realistically, true interoperability will take time to perfect, and these efforts require strong industry collaboration among regulators, administrators, scientists, and, just as importantly, physicians. With more EHRs embracing Fast Healthcare Interoperability Resources (FHIR, a standard for health care data to facilitate exchange), there is progress in the right direction, but there remains enough variation to prevent rapid or reliable interoperability.14 Simultaneous with these infrastructure challenges are a variety of additional technical challenges related to model deployment. Until recently, none of the major EHR companies supported machine learning out of the box. Ideally, there would be a machine-learning platform within the EHR to easily integrate existing scripts written with common machine-learning tools in a manner that would not affect the real-time performance of the EHR. The analogy would be the app store for a cell phone. Developers are able to develop software that runs on a third party’s platform (the cell phone manufacturer) in a reliable and predictable way. The cell phone maker publishes clear standards and interoperability rules but has little to do with the actual app development. To date, despite the hype surrounding machine learning, none of the current EHR platforms offer services anywhere near as robust as those that can be created outside EHR platforms. This gap requires developers to export the data from the EHR into another platform to run their models, increasing the potential failure points and requiring significant additional infrastructure. Deployment of machine-learning models has nontechnical hurdles. Policies need to be in place to ensure appropriate oversight, specifically delineating who is responsible for the data, model upkeep, and any actionable consequences of model results. Stakeholders must be committed to guarantee sustainable implementation, as in the instituting of appropriate patient privacy and consent policies.15 In addition, there are a number of costs associated with development, whether they be financial costs or time costs. Neither of these can be underappreciated or reduced in an attempt to economize. Health care is a highly complex system, and the stakes are high in the event of model failures. “Moving fast and breaking things,16” as has been described in Silicon Valley, is not a viable option in medicine. If a global positioning system (GPS) app suggests the wrong route to home, the consequence is a longer commute. If a model fails in health care, the consequences are life threatening. TOWARD THE FUTURE So what does the future hold? While the exact timeline or path to adoption remains ill defined, machine learning will certainly become part of, and alter, medicine. Despite the challenges in predicting the future, answering 3 key questions may be helpful in understanding the ways that machine learning will impact anesthesiology. What clinical challenges most lend themselves to machine learning–based implementations? Those most successfully approached can be broadly classified into 2 groups: questions assessing clinical treatment strategies and those assessing clinical risk. There are clinical scenarios where practice is unclear at the time. For example, several studies have shown that machine-learning models are capable of discriminating which septic patients are most likely to be volume responsive.17 Similarly, several recent studies have shown that models can be constructed to help predict perioperative risk,18 something that clinicians find notoriously hard to quantify. Thus, there is significant potential for models to help clinicians identify at-risk patients and ultimately quantify what is currently not quantifiable. How might we overcome obstacles? Historical lessons from other industries may be of use. In the 1990s, as computers were becoming more relevant, implementation was plagued with issues related to a lack of standardization and interoperability. Computer companies attempted to use proprietary cables, closed software formats, and other similar devices to ensure customer loyalty. However, over time, a combination of innovation and regulation served to create standards that allowed the computer industry to make hardware and operating systems, mostly manufacturer agnostic. The aforementioned FHIR standards are likely the early beginnings of these changes in medicine. Implementation and workflow are likely the places where clinicians must be most actively involved in machine-learning applications. Anesthesia has historically been entrenched in informatics,19 and the anesthesiologist will have continued involvement in advocating for patients and ensuring that machine-learning implementations augment rather than complicate clinical workflow. These efforts, including reducing the increasing requirements for data entry, will be critical for successful advancement and adoption of artificial intelligence in health care. The last question is, what might the future of health care look like? It is conceivable that, in the coming decades, many of the repetitive tasks we perform today will become automated. Others have described target-controlled infusion systems whereby an anesthesiologist sets blood pressure goals and the pressor dose is automatically adjusted to keep the pressure within the designated zone. Machine learning has the ability to take this a step further, determining patient-specific blood pressure goals and extrapolating results from previous cases to determine which pressor to use. However, at its core, the practice of medicine will remain largely unchanged. Medicine involves speaking to patients and helping them in what are often the most stressful times of their lives. These higher cognitive tasks, such as understanding a patient’s goals and weighing competing priorities, are unlikely to ever be outsourced to a machine and will likely become a bigger part of the clinical effort of physicians. It is for this reason the term “artificial intelligence” is increasingly being recognized as a misnomer. The future is likely one of augmented intelligence, where computers become indispensable tools that help us care for our patients, and the physician can devote more of their efforts on patient care. DISCLOSURES Name: Ira S. Hofer, MD. Contribution: This author helped write the manuscript. Conflicts of Interest: I. S. Hofer receives funding from Merck Pharmaceuticals. He is also the founder and president of Clarity Healthcare Analytics and has a copyright on software that assists with data extraction from the EHR. Name: Michael Burns, MD, PhD. Contribution: This author helped write the manuscript. Conflicts of Interest: None. Name: Samir Kendale, MD. Contribution: This author helped write the manuscript. Conflicts of Interest: None. Name: Jonathan P. Wanderer, MD, MPhil, FASA, FAMIA. Contribution: This author helped write the manuscript. Conflicts of Interest: None. This manuscript was handled by: Maxime Cannesson, MD, PhD.

Ähnliche Arbeiten

Autoren

Institutionen

Themen

Artificial Intelligence in Healthcare and EducationMachine Learning in HealthcareElectronic Health Records Systems
Volltext beim Verlag öffnen