OpenAlex · Aktualisierung stündlich · Letzte Aktualisierung: 13.05.2026, 08:47

Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.

Vision-language large learning model, GPT4V, accurately classifies the Boston Bowel Preparation Scale score

2025·5 Zitationen·BMJ Open GastroenterologyOpen Access
Volltext beim Verlag öffnen

5

Zitationen

16

Autoren

2025

Jahr

Abstract

INTRODUCTION: Large learning models (LLMs) such as GPT are advanced artificial intelligence (AI) models. Originally developed for natural language processing, they have been adapted for multi-modal tasks with vision-language input. One clinically relevant task is scoring the Boston Bowel Preparation Scale (BBPS). While traditional AI techniques use large amounts of data for training, we hypothesise that vision-language LLM can perform this task with fewer examples. METHODS: (GIE 2009). Performance was tested on the HyperKvasir dataset, an open dataset for automated BBPS grading. RESULTS: Of 1794 images, GPT4V returned valid results for 1772 (98%). It had an accuracy of 0.84 for two-class classification (BBPS 0-1 vs 2-3) and 0.74 for four-class classification (BBPS 0, 1, 2, 3). Macro-averaged F1 scores were 0.81 and 0.63, respectively. Qualitatively, most errors arose from misclassification of BBPS 1 as 2. These results compare favourably with current methods using large amounts of training data, which achieve an accuracy in the range of 0.8-0.9. CONCLUSION: This study provides proof-of-concept that a vision-language LLM is able to perform BBPS classification accurately, without large training datasets. This represents a paradigm shift in AI classification methods in medicine, where many diseases lack sufficient data to train traditional AI models. An LLM with appropriate examples may be used in such cases.

Ähnliche Arbeiten