OpenAlex · Aktualisierung stündlich · Letzte Aktualisierung: 22.04.2026, 07:33

Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.

Fine-tuned Language Models are Continual Learners

2022·7 Zitationen·arXiv (Cornell University)Open Access
Volltext beim Verlag öffnen

7

Zitationen

3

Autoren

2022

Jahr

Abstract

Recent work on large language models relies on the intuition that most natural language processing tasks can be described via natural language instructions. Language models trained on these instructions show strong zero-shot performance on several standard datasets. However, these models even though impressive still perform poorly on a wide range of tasks outside of their respective training and evaluation sets. To address this limitation, we argue that a model should be able to keep extending its knowledge and abilities, without forgetting previous skills. In spite of the limited success of Continual Learning we show that Language Models can be continual learners. We empirically investigate the reason for this success and conclude that Continual Learning emerges from self-supervision pre-training. Our resulting model Continual-T0 (CT0) is able to learn diverse new tasks, while still maintaining good performance on previous tasks, spanning remarkably through 70 datasets in total. Finally, we show that CT0 is able to combine instructions in ways it was never trained for, demonstrating some compositionality.

Ähnliche Arbeiten

Autoren

Themen

Topic ModelingDomain Adaptation and Few-Shot LearningArtificial Intelligence in Healthcare and Education
Volltext beim Verlag öffnen