OpenAlex · Aktualisierung stündlich · Letzte Aktualisierung: 04.05.2026, 01:58

Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.

HUR Architecture with HurNet: Revolutionizing Language Model Training through Single-Step Weight Optimization

2026·0 Zitationen·Zenodo (CERN European Organization for Nuclear Research)Open Access
Volltext beim Verlag öffnen

0

Zitationen

1

Autoren

2026

Jahr

Abstract

This paper introduces HurNet, a novel neural network architecture that revolutionizes the training of large language models by replacing iterative backpropagation with direct, single-step division-based computations. By intelligently initializing weights using pseudo-inverse or direct division methods, HurNet achieves optimal parameter configurations more rapidly than traditional gradient descent approaches. The proposed Hur Architecture integrates HurNet with Transformer-based models, enabling efficient training on modest hardware with reduced computational costs. Key features include automatic hyperparameter configuration, customizable fine-tuning modalities, mixture of experts integration, model merging, and transfer learning capabilities. Additionally, Hur Architecture supports infinite context windows through a sophisticated embedding synthesis mechanism formalized via probabilistic concatenation operators. Empirical evaluations demonstrate significant speedups, with Hur models outperforming conventional GPT architectures by up to 582% in fine-tuning tasks. The resulting models are serialized in the .hurlm format, facilitating deployment across resource-constrained environments.

Ähnliche Arbeiten

Autoren

Institutionen

Themen

Machine Learning and Data ClassificationAdvanced Neural Network ApplicationsArtificial Intelligence in Healthcare and Education
Volltext beim Verlag öffnen