Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
MedVisionLlama: Leveraging Pre-Trained Large Language Model Layers to Enhance Medical Image Segmentation
1
Zitationen
4
Autoren
2025
Jahr
Abstract
Medical image segmentation plays a key role in healthcare, enabling accurate diagnosis and treatment planning. Vision Transformers (ViTs) show strong potential for segmentation tasks, but their dependence on large datasets limits practical usage in clinical settings. This study explores whether integrating pre-trained Large Language Models (LLMs) with ViT-based segmentation models can enhance feature refinement and improve performance in data-constrained environments. We introduce MedVisionLlama, which combines ViT encoders with pre-trained Llama weights and applies Low-Rank Adaptation (LoRA) for fine-tuning in 3D medical image segmentation. Evaluated on the Medical Segmentation Decathlon dataset, the model consistently outperformed a standard ViT, showing improved generalization across MRI and CT modalities. It maintained stable segmentation quality even with limited training data and across varied anatomical structures. Activation maps revealed sharper and more stable attention to relevant regions. Ablation studies confirmed that the performance gains stemmed from LLM-based feature refinement rather than increased model complexity. MedVisionLlama offers a scalable and data-efficient solution for medical image segmentation. Source code and implementation are available at: https://github.com/AS-Lab/Marthi-etal-2025-MedVisionLlama-Pre-Trained-LLM-Layers-to-Enhance-Medical-Image-Segmentation.
Ähnliche Arbeiten
MizAR 60 for Mizar 50
2023 · 74.703 Zit.
ImageNet: A large-scale hierarchical image database
2009 · 60.749 Zit.
Microsoft COCO: Common Objects in Context
2014 · 41.350 Zit.
Fully convolutional networks for semantic segmentation
2015 · 36.453 Zit.
Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization
2017 · 20.561 Zit.