OpenAlex · Aktualisierung stündlich · Letzte Aktualisierung: 05.04.2026, 06:36

Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.

MedVisionLlama: Leveraging Pre-Trained Large Language Model Layers to Enhance Medical Image Segmentation

2025·1 Zitationen
Volltext beim Verlag öffnen

1

Zitationen

4

Autoren

2025

Jahr

Abstract

Medical image segmentation plays a key role in healthcare, enabling accurate diagnosis and treatment planning. Vision Transformers (ViTs) show strong potential for segmentation tasks, but their dependence on large datasets limits practical usage in clinical settings. This study explores whether integrating pre-trained Large Language Models (LLMs) with ViT-based segmentation models can enhance feature refinement and improve performance in data-constrained environments. We introduce MedVisionLlama, which combines ViT encoders with pre-trained Llama weights and applies Low-Rank Adaptation (LoRA) for fine-tuning in 3D medical image segmentation. Evaluated on the Medical Segmentation Decathlon dataset, the model consistently outperformed a standard ViT, showing improved generalization across MRI and CT modalities. It maintained stable segmentation quality even with limited training data and across varied anatomical structures. Activation maps revealed sharper and more stable attention to relevant regions. Ablation studies confirmed that the performance gains stemmed from LLM-based feature refinement rather than increased model complexity. MedVisionLlama offers a scalable and data-efficient solution for medical image segmentation. Source code and implementation are available at: https://github.com/AS-Lab/Marthi-etal-2025-MedVisionLlama-Pre-Trained-LLM-Layers-to-Enhance-Medical-Image-Segmentation.

Ähnliche Arbeiten

Autoren

Institutionen

Themen

Multimodal Machine Learning ApplicationsArtificial Intelligence in Healthcare and EducationCOVID-19 diagnosis using AI
Volltext beim Verlag öffnen