Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Large Language Models for Endodontic Diagnosis: A Comparative Study Against an Expert Reference Standard
0
Zitationen
3
Autoren
2025
Jahr
Abstract
Accurate diagnosis is the cornerstone of successful endodontic care, yet it often remains a challenge in daily practice. Artificial intelligence (AI), particularly large language models (LLMs), is being explored as a possible support tool for clinicians. In this study, we compared the diagnostic ability of three LLMs: ChatGPT-5, Gemini, and Perplexity with the judgment of an experienced endodontist across 40 anonymized clinical cases. ChatGPT-5 reached perfect agreement with the expert (100% accuracy, κ = 1.00). Gemini (97.5% accuracy, κ = 0.95) and Perplexity (92.5% accuracy, κ = 0.85) also performed well but showed different error patterns: Gemini made one false positive, while Perplexity missed three positive cases. Overall, ChatGPT-5 and Gemini showed the best sensitivity, and ChatGPT-5 and Perplexity maintained full specificity. These results show that the latest LLMs can approach expert-level diagnostic performance, but further testing is needed before they can be relied upon in everyday clinical practice.
Ähnliche Arbeiten
Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI
2019 · 8.239 Zit.
Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead
2019 · 8.095 Zit.
High-performance medicine: the convergence of human and artificial intelligence
2018 · 7.463 Zit.
Proceedings of the 19th International Joint Conference on Artificial Intelligence
2005 · 5.776 Zit.
Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI)
2018 · 5.428 Zit.