OpenAlex · Aktualisierung stündlich · Letzte Aktualisierung: 15.03.2026, 16:14

Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.

Optimizing Large Language Models in Distributed Environments: A Holistic Approach to Efficiency, Ethics, and Governance

2025·0 Zitationen·International Journal of Computational Intelligence SystemsOpen Access
Volltext beim Verlag öffnen

0

Zitationen

6

Autoren

2025

Jahr

Abstract

This paper introduces a holistic and scalable framework for optimizing Large Language Models (LLMs) in distributed environments, addressing three critical challenges: computational efficiency, ethical fairness, and governance. As LLMs scale, issues, such as excessive resource consumption, fairness violations, and limited transparency, hinder their broader deployment in real-world applications. We propose a novel three-tier architecture that integrates topology-aware parallelism, communication-efficient gradient aggregation, and memory-aware rematerialization. Our implementation reduces training time by 38% and memory usage by 42% on a 512-GPU A100 cluster, without compromising accuracy. To promote fairness, we incorporate a real-time adversarial debiasing module that reduces demographic AUC gaps by over 60% across gender, ethnicity, and religion. For model interpretability, we introduce a symbolic explainability engine that converts attention weights into transparent rule-based explanations, achieving 89.2% user satisfaction and outperforming Grad-CAM and vanilla attention. Furthermore, a lightweight governance layer aligned with ISO/IEC 27001 and ISO/IEC 23894 standards ensures traceability, audit logging, and policy enforcement throughout the model lifecycle. We validate our framework across diverse datasets, including C4, WikiText-103, RealNews, and BookCorpus, demonstrating low-latency drift and consistent fairness across domains. Comparative benchmarks against DeepSpeed, FairScale, and Megatron-LM show superior throughput, energy efficiency, and transparency. This work advances the foundation for ethical, efficient, and regulation-compliant LLM deployment at scale.

Ähnliche Arbeiten