Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Optimizing Large Language Models in Distributed Environments: A Holistic Approach to Efficiency, Ethics, and Governance
0
Zitationen
6
Autoren
2025
Jahr
Abstract
This paper introduces a holistic and scalable framework for optimizing Large Language Models (LLMs) in distributed environments, addressing three critical challenges: computational efficiency, ethical fairness, and governance. As LLMs scale, issues, such as excessive resource consumption, fairness violations, and limited transparency, hinder their broader deployment in real-world applications. We propose a novel three-tier architecture that integrates topology-aware parallelism, communication-efficient gradient aggregation, and memory-aware rematerialization. Our implementation reduces training time by 38% and memory usage by 42% on a 512-GPU A100 cluster, without compromising accuracy. To promote fairness, we incorporate a real-time adversarial debiasing module that reduces demographic AUC gaps by over 60% across gender, ethnicity, and religion. For model interpretability, we introduce a symbolic explainability engine that converts attention weights into transparent rule-based explanations, achieving 89.2% user satisfaction and outperforming Grad-CAM and vanilla attention. Furthermore, a lightweight governance layer aligned with ISO/IEC 27001 and ISO/IEC 23894 standards ensures traceability, audit logging, and policy enforcement throughout the model lifecycle. We validate our framework across diverse datasets, including C4, WikiText-103, RealNews, and BookCorpus, demonstrating low-latency drift and consistent fairness across domains. Comparative benchmarks against DeepSpeed, FairScale, and Megatron-LM show superior throughput, energy efficiency, and transparency. This work advances the foundation for ethical, efficient, and regulation-compliant LLM deployment at scale.
Ähnliche Arbeiten
Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI
2019 · 8.239 Zit.
Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead
2019 · 8.095 Zit.
High-performance medicine: the convergence of human and artificial intelligence
2018 · 7.463 Zit.
Proceedings of the 19th International Joint Conference on Artificial Intelligence
2005 · 5.776 Zit.
Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI)
2018 · 5.428 Zit.