Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Secure yet fragile: adversarial vulnerabilities of federated vision–language models in medical AI
0
Zitationen
8
Autoren
2026
Jahr
Abstract
Vision-Language Models (VLMs) enable powerful multimodal reasoning for medical image analysis, while federated learning allows collaborative training across institutions without sharing patient data. However, the adversarial robustness of federated medical VLMs remains largely unexplored. This work systematically evaluates the vulnerability of CLIP-based VLMs trained with four federated optimization strategies, FedAvg, FedProx, FedPer, and FedBN, on multiple medical datasets. We assess robustness under FGSM, PGD, BIM, and MI-FGSM attacks at varying strengths and show that client-level adversarial perturbations propagate through federated aggregation, causing severe accuracy degradation and high attack success rates, specially under iterative attacks. We further benchmark two training-free test-time defenses, Test-Time Counter-Attack (TTC) and CLIPure, and demonstrate that both mitigate adversarial effects, with CLIPure providing more consistent improvements across datasets and attack intensities. These results highlight fundamental robustness limitations of federated medical VLMs and underscore the need for effective defense mechanisms in distributed clinical deployments.
Ähnliche Arbeiten
Rethinking the Inception Architecture for Computer Vision
2016 · 30.537 Zit.
MobileNetV2: Inverted Residuals and Linear Bottlenecks
2018 · 24.719 Zit.
CBAM: Convolutional Block Attention Module
2018 · 21.614 Zit.
An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale
2020 · 21.412 Zit.
Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification
2015 · 18.606 Zit.