Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Automated ACR TI-RADS Classification of Thyroid Nodules from Narrative Ultrasound Reports Using a Fine-Tuned Open-Source Language Model: A Reproducible and Low-Resource Framework
0
Zitationen
8
Autoren
2025
Jahr
Abstract
<title>Abstract</title> Background: Manual ACR TI-RADS classification from narrative ultrasound reports is a key component of thyroid nodule risk stratification but is laborious and subject to inter-observer variability. While Large Language Models (LLMs) offer potential solutions, existing approaches often rely on proprietary models or require extensive computational resources, limiting widespread adoption. This study aimed to develop and validate a reproducible, low-resource framework using a fine-tuned open-source LLM to automate this task. Methods: This retrospective study utilized a dataset of 1,850 de-identified thyroid ultrasound reports from a primary single center. The reports were annotated by radiologists to establish a ground truth. An open-source 7-billion parameter model (Qwen1.5-7B) was fine-tuned on a training set (n=1,480) using Low-Rank Adaptation (LoRA) on a single consumer-grade GPU. The model's performance was evaluated on a hold-out internal test set (n=370) and a separate external validation set (n=210) from another institution. Results: On the internal test set, the fine-tuned model achieved an overall accuracy of 93.0% and a macro-averaged F1-score of 0.950. On the external validation set, it maintained robust performance with an accuracy of 88.6% and a macro F1-score of 0.891, demonstrating strong generalizability. It significantly outperformed both a zero-shot LLM baseline and a traditional machine learning model (TF-IDF with SVM) on both datasets. Conclusions: Fine-tuning an accessible, open-source language model on local, consumer-grade hardware is an effective and resource-efficient strategy for automating ACR TI-RADS classification from narrative reports. This approach offers a practical and generalizable blueprint for healthcare institutions to develop bespoke AI tools, potentially enhancing workflow efficiency and diagnostic consistency while preserving data privacy.
Ähnliche Arbeiten
Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI
2019 · 8.245 Zit.
Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead
2019 · 8.100 Zit.
High-performance medicine: the convergence of human and artificial intelligence
2018 · 7.466 Zit.
Proceedings of the 19th International Joint Conference on Artificial Intelligence
2005 · 5.776 Zit.
Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI)
2018 · 5.429 Zit.