OpenAlex · Aktualisierung stündlich · Letzte Aktualisierung: 01.05.2026, 10:42

Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.

An LLM-Enabled Multimodal Agentic AI Framework for the Medical Internet of Things (MIoT)

2026·0 Zitationen·ACM Transactions on Multimedia Computing Communications and Applications
Volltext beim Verlag öffnen

0

Zitationen

9

Autoren

2026

Jahr

Abstract

The integration of large language models (LLM) with multimodal agentic artificial intelligence (AI) within the medical Internet of Things (MIoT) ecosystem is redefining modern healthcare intelligence. This convergence enables continuous patient observation, adaptive clinical decision-making, and context-aware interaction between humans and machines across various biomedical data modalities. Healthcare systems generate a wide range of multimodal data, including textual records such as EHRs, prescriptions, and pathology notes; medical imagery such as CT, MRI, fundus, and radiographs; spoken data from consultations and transcriptions; video streams for rehabilitation and physiotherapy monitoring; and sensor readings such as ECG, SpO \({}_{2}\) , and glucose levels. Conventional unimodal algorithms fall short in interpreting this diversity, whereas LLM-augmented agentic frameworks fuse and reason over these heterogeneous sources, grounding their outputs in medical ontologies and coordinating task-specific agents to enhance real-world clinical workflows. This paper presents a comprehensive overview of multimodal agentic AI powered by LLM for MIoT-enabled healthcare systems. Introduces a six-dimensional unified taxonomy that covers multimodal input channels, fusion mechanisms, core LLM reasoning capabilities, agentic coordination models, computational deployment layers, and ethical governance frameworks. To contextualize this taxonomy, the discussion includes a Virtual Hospital case study centered on cancer that demonstrates how multimodal signals such as imaging, genomics, patient dialogues, and clinical updates integrate through intelligent agents to enable personalized diagnosis, automated documentation, home rehabilitation, and rapid intervention in emergencies. The survey also consolidates current progress on datasets, benchmarks, and evaluation protocols for AI in multimodal and agentic healthcare. The survey identifies critical research gaps, such as the lack of longitudinal multimodal datasets, standardized evaluation frameworks for multi-agent reasoning, and reliable methods to assess trustworthiness in clinical AI. Furthermore, it examines security and compliance issues such as adversarial manipulation, data leakage, and accountability across distributed agent networks, and it proposes countermeasures through federated data governance, secure MCP-oriented orchestration, and privacy-aware edge deployment strategies. By situating recent advances within the Virtual Hospital paradigm and oncology workflows, this study provides a systematic foundation for developing scalable, secure, and ethically aligned multimodal agentic systems based on LLMs, guiding the next generation of intelligent MIoT-driven healthcare ecosystems

Ähnliche Arbeiten