Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
ChatGPT-CARE: a Superior Decision Support Tool Enhancing ChatGPT with Clinical Practice Guidelines
8
Zitationen
5
Autoren
2023
Jahr
Abstract
Abstract ChatGPT has gained remarkable traction since its inception in November 2022. However, it faces limitations in generating inaccurate responses, ignoring existing guidelines, and lacking reasoning when applied in clinical settings. This study introduces ChatGPT-CARE, a tool that integrates clinical practice guidelines with ChatGPT, focusing on COVID-19 outpatient treatment decisions. By employing in-context learning and chain-of-thought prompting techniques, ChatGPT-CARE enhances original ChatGPT’s clinical decision support and reasoning capabilities. We created three categories of various descriptions of patients seeking COVID-19 treatment to evaluate the proposed tool, and asked two physicians specialized in pulmonary disease and critical care to assess the responses for accuracy, hallucination, and clarity. The results indicate that ChatGPT-CARE offers increased accuracy and clarity, with moderate hallucination, compared to the original ChatGPT. The proposal ChatGPT-CARE could be a viable AI-driven clinical decision support tool superior to ChatGPT, with potential applications beyond COVID-19 treatment decision support.
Ähnliche Arbeiten
Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI
2019 · 8.200 Zit.
Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead
2019 · 8.051 Zit.
High-performance medicine: the convergence of human and artificial intelligence
2018 · 7.416 Zit.
Proceedings of the 19th International Joint Conference on Artificial Intelligence
2005 · 5.776 Zit.
Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI)
2018 · 5.410 Zit.