Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Federated Instruction Tuning with DeepSeek: Towards Scalable and Private LLM Adaptation
0
Zitationen
3
Autoren
2025
Jahr
Abstract
Recent advances in large language models (LLMs) have enabled a wide range of AI-driven applications, yet their deployment in sensitive domains such as healthcare remains constrained by stringent data privacy requirements. To address this challenge, we adopt a federated learning framework for cross-institutional collaborative training of the DeepSeek-R1-Distill-Qwen-1.5B model, thereby avoiding centralized data aggregation. In this setup, each institution trains the model locally on its data, and updates are aggregated using Federated Averaging (FedAvg) to construct a global model while ensuring patient information never leaves its source. Experiments conducted on the medical-o1-reasoning-SFT dataset demonstrate that our federated approach achieves 56.9% accuracy, closely matching the performance of centralized training while preserving data confidentiality. Furthermore, the model maintains stable performance under a non-IID manual split with per-client subsampling, underscoring the robustness of our method.
Ähnliche Arbeiten
k-ANONYMITY: A MODEL FOR PROTECTING PRIVACY
2002 · 8.402 Zit.
Calibrating Noise to Sensitivity in Private Data Analysis
2006 · 6.895 Zit.
Deep Learning with Differential Privacy
2016 · 5.629 Zit.
Communication-Efficient Learning of Deep Networks from Decentralized\n Data
2016 · 5.595 Zit.
Federated Machine Learning
2019 · 5.581 Zit.