Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
LLLMs: A Data-Driven Survey of Evolving Research on Limitations of Large Language Models
1
Zitationen
6
Autoren
2026
Jahr
Abstract
Large language model (LLM) research has grown rapidly, along with increasing concern about their limitations. In this survey, we conduct a data-driven, semi-automated review of research on limitations of LLMs ( LLLMs ) from 2022 to early 2025 using a bottom-up approach. From a corpus of 250,000 ACL and arXiv papers, we identify 14,648 relevant papers using keyword filtering, LLM-based classification, validated against expert labels, and topic clustering (via two approaches, HDBSCAN+BERTopic and LlooM). We find that the share of LLM-related papers increases over fivefold in ACL and nearly eightfold in arXiv between 2022 and 2025. Since 2022, LLLMs research grows even faster, reaching over 30% of LLM papers by 2025. Reasoning remains the most studied limitation, followed by generalization , hallucination , bias , and security . The distribution of topics in the ACL dataset stays relatively stable over time, while arXiv shifts toward security risks , alignment , hallucinations , knowledge editing , and multimodality . We offer a quantitative view of trends in LLLMs research and release a dataset of annotated abstracts and a validated methodology, available at: github.com/a-kostikova/LLLMs-Survey.
Ähnliche Arbeiten
2019 · 32.085 Zit.
Techniques to Identify Themes
2003 · 5.408 Zit.
Answering the Call for a Standard Reliability Measure for Coding Data
2007 · 4.113 Zit.
Basic Content Analysis
1990 · 4.045 Zit.
Text as Data: The Promise and Pitfalls of Automatic Content Analysis Methods for Political Texts
2013 · 3.113 Zit.