Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Its All About Context
2
Zitationen
3
Autoren
2024
Jahr
Abstract
This paper presents an innovative approach to enhancing context understanding in Generative Pretrained Models (GPTs), a critical step towards achieving Artificial General Intelligence (AGI). While GPTs have significantly advanced natural language processing, their understanding of context remains predominantly limited to linguistic structure. To address this limitation, we introduce a novel layer in the transformer model architecture that computes context weights, integrating both immediate and temporally decaying influences of past tokens. This layer is strategically positioned after the self-attention and before the feed-forward layers, enabling a more nuanced interpretation of sequential language data. Our approach involves the formulation of a decaying temporal factor, which allows the model to consider not only the immediate relevance of tokens but also their historical context. This factor is dynamically adjustable, offering a sophisticated method of context weighting that considers both current and extended contexts. The integration of this context weight into the self-attention mechanism enhances the model's capacity for a deeper, more accurate understanding of language, pushing the boundaries of current AI capabilities towards a system that mirrors human-like intelligence. Our experimental results demonstrate the efficacy of this approach, showing its potential to significantly contribute to the development of AGI.
Ähnliche Arbeiten
Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI
2019 · 8.200 Zit.
Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead
2019 · 8.051 Zit.
High-performance medicine: the convergence of human and artificial intelligence
2018 · 7.416 Zit.
Proceedings of the 19th International Joint Conference on Artificial Intelligence
2005 · 5.776 Zit.
Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI)
2018 · 5.410 Zit.