Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
<i>Editorial Commentary</i> : Shifting From Redundancy to Rigor in Orthopaedic Large Language Model Research
1
Zitationen
3
Autoren
2025
Jahr
Abstract
Large language model (LLM) research in musculoskeletal medicine is growing rapidly, but much of the literature remains methodologically weak and highly repetitive. To address this, the orthopaedic LLM research community needs a shared benchmarking infrastructure/framework to evaluate models on clinically grounded tasks using fixed-prompt templates and transparent scoring. Drawing on established LLM benchmarking practices, such a framework would enable reproducibility, discourage cherry-picking, and promote meaningful innovation. Like surgical registries in orthopaedics, open LLM benchmarks can clarify performance, guide adoption, and ensure that progress is both measurable and clinically relevant.
Ähnliche Arbeiten
Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI
2019 · 8.260 Zit.
Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead
2019 · 8.116 Zit.
High-performance medicine: the convergence of human and artificial intelligence
2018 · 7.493 Zit.
Proceedings of the 19th International Joint Conference on Artificial Intelligence
2005 · 5.776 Zit.
Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI)
2018 · 5.438 Zit.