OpenAlex · Aktualisierung stündlich · Letzte Aktualisierung: 02.05.2026, 10:33

Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.

Self-Attention and Transformers: Driving the Evolution of Large Language Models

2023·36 Zitationen
Volltext beim Verlag öffnen

36

Zitationen

6

Autoren

2023

Jahr

Abstract

Transformers, originally introduced for machine translation, and built upon the Self-Attention mechanism, have undergone a remarkable evolution, establishing themselves as the bedrock of large language models (LLMs). Their unparalleled capacity to model intricate relationships and capture extensive dependencies within sequences has propelled their prominence. This article, presented in a popular science format, serves as an introduction to the transformer architecture, elucidating its innovative structure that enables efficient processing of long sequences and capturing dependencies over extended distances. We believe that this resource will prove valuable to college students or youth researchers aspiring to delve into the study and research of modern Artificial Intelligence (AI) domains.

Ähnliche Arbeiten

Autoren

Institutionen

Themen

Topic ModelingArtificial Intelligence in Healthcare and EducationNatural Language Processing Techniques
Volltext beim Verlag öffnen