Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
An exploratory study on the potential of ChatGPT as an AI-assisted diagnostic tool for visceral leishmaniasis
2
Zitationen
16
Autoren
2024
Jahr
Abstract
Visceral leishmaniasis (VL) is a severe parasitic disease that poses significant diagnostic challenges due to its complex presentation and the necessity for comprehensive diagnostic methods. This exploratory study investigates the potential of Chat Generative Pre-trained Transformer (ChatGPT)/GPT-4, an artificial intelligence (AI) chatbot, in assisting the diagnostic process for VL. We evaluated the diagnostic accuracy of ChatGPT/GPT-4 in generating differential diagnosis lists for eight clinical vignette cases of VL, authored by a Brazilian infectious disease doctor. Our findings reveal that ChatGPT/GPT-4 included VL in the top five differential diagnoses in 75% of the cases (95% confidence interval [CI]: 40.1 – 93.7%) and identified VL as the top diagnosis in 50% of the cases (95% CI: 30.3 – 86.5%). These results underscore the high potential of ChatGPT/GPT-4 as an AI-assisted diagnostic tool, which is capable of providing accurate differential diagnoses and assisting healthcare professionals in resource-limited settings. The study highlights the broader applicability of AI chatbots in medical diagnostics, not only for common conditions but also for specialized and less prevalent diseases like VL. By integrating AI tools into the diagnostic workflow, healthcare providers can enhance their diagnostic accuracy and efficiency, ultimately improving patient outcomes. This research contributes to the growing body of evidence supporting the utility of AI in healthcare and underscores the need for further studies to validate these findings across larger and more diverse clinical scenarios.
Ähnliche Arbeiten
Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI
2019 · 8.324 Zit.
Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead
2019 · 8.189 Zit.
High-performance medicine: the convergence of human and artificial intelligence
2018 · 7.588 Zit.
Proceedings of the 19th International Joint Conference on Artificial Intelligence
2005 · 5.776 Zit.
Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI)
2018 · 5.470 Zit.
Autoren
- Paulo Adriano Schwingel
- Dino Schwingel
- Samuel Ricarte de Aquino
- Aline Rafaela Soares da Silva
- Pedro Paulo Ramos da Silva
- R Pereira
- Daniela Conceição Gomes Gonçalves e Silva
- Amanda Alves Marcelino da Silva
- Flávia Emília Cavalcante Valença Fernandes
- Maria Jacqueline Silva Ribeiro
- Paulo Ditarso Maciel Júnior
- Paulo Gustavo Serafim de Carvalho
- Ricardo Kenji Shiosaki
- Raquel M. Gonçalves
- Bruno Bavaresco Gambassi
- Paula Andreatta Maduro