Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
A Comprehensive Survey in LLM(-Agent) Full Stack Safety: Data, Training and Deployment
0
Zitationen
103
Autoren
2025
Jahr
Abstract
The remarkable success of Large Language Models (LLMs) has illuminated a promising pathway toward achieving Artificial General Intelligence for both academic and industrial communities, owing to their unprecedented performance across various applications. As LLMs continue to gain prominence in both research and commercial domains, their security and safety implications have become a growing concern, not only for researchers and corporations but also for every nation. Currently, existing surveys on LLM safety primarily focus on specific stages of the LLM lifecycle, e.g., deployment phase or fine-tuning phase, lacking a comprehensive understanding of the entire "lifechain" of LLMs. To address this gap, this paper introduces, for the first time, the concept of "full-stack" safety to systematically consider safety issues throughout the entire process of LLM training, deployment, and eventual commercialization. Compared to the off-the-shelf LLM safety surveys, our work demonstrates several distinctive advantages: (I) Comprehensive Perspective. We define the complete LLM lifecycle as encompassing data preparation, pre-training, post-training, deployment and final commercialization. To our knowledge, this represents the first safety survey to encompass the entire lifecycle of LLMs. (II) Extensive Literature Support. Our research is grounded in an exhaustive review of over 800+ papers, ensuring comprehensive coverage and systematic organization of security issues within a more holistic understanding. (III) Unique Insights. Through systematic literature analysis, we have developed reliable roadmaps and perspectives for each chapter. Our work identifies promising research directions, including safety in data generation, alignment techniques, model editing, and LLM-based agent systems. These insights provide valuable guidance for researchers pursuing future work in this field.
Ähnliche Arbeiten
Rethinking the Inception Architecture for Computer Vision
2016 · 30.338 Zit.
MobileNetV2: Inverted Residuals and Linear Bottlenecks
2018 · 24.418 Zit.
An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale
2020 · 21.303 Zit.
CBAM: Convolutional Block Attention Module
2018 · 21.301 Zit.
Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification
2015 · 18.499 Zit.
Autoren
- Kun Wang
- Guibin Zhang
- Zhenhong Zhou
- Jiahao Wu
- Miao Yu
- Shiqian Zhao
- Chenlong Yin
- Fu J
- Y. H. Yan
- Hao Luo
- Liang Lin
- Zhihao Xu
- Hai Lu
- Xue Cao
- Xinyun Zhou
- Weifei Jin
- Fanci Meng
- Shicheng Xu
- Junyuan Mao
- Yu Wang
- Hao Wu
- Minghe Wang
- Fan Zhang
- Junfeng Fang
- Wenjie Qu
- Yue Liu
- Chengwei Liu
- Yifan Zhang
- Qiankun Li
- Chunbao Guo
- Yalan Qin
- Zhaoxin Fan
- Kai Wang
- Yi Ding
- Donghai Hong
- Jiaming Ji
- Yingxin Lai
- Zitong Yu
- Xinfeng Li
- Yifan Jiang
- Yanhui Li
- Xinyu Deng
- Junlin Wu
- Dongxia Wang
- Yihao Huang
- Yufei Guo
- Jen-tse Huang
- Qiufeng Wang
- Xiaolong Jin
- Wenxuan Wang
- Dongrui Liu
- Yanwei Yue
- Wenke Huang
- Guancheng Wan
- Heng Chang
- Tianlin Li
- Yi Yu
- Chenghao Li
- Jiawei Li
- Lei Bai
- Jie Zhang
- Qing Guo
- Jingyi Wang
- Tianlong Chen
- Jia Zhou
- Xiaojun Jia
- Weisong Sun
- Cong Wu
- Jing Chen
- Xuming Hu
- Yiming Li
- Xiao Wang
- Ningyu Zhang
- Luu Anh Tuan
- Guowen Xu
- Jiaheng Zhang
- Tianwei Zhang
- Xingjun Ma
- Jindong Gu
- Liang Pang
- Xuan Wang
- Bo An
- Jun Sun
- Mohit Bansal
- Shirui Pan
- Lingjuan Lyu
- Yuval Elovici
- Bhavya Kailkhura
- Yaodong Yang
- Hongwei Li
- Wenyuan Xu
- Yizhou Sun
- Wei Wang
- Qing Li
- Ke Tang
- Yu-Gang Jiang
- Felix Juefei-Xu
- Wei-Hua Lin
- Xiaofeng Wang
- Dacheng Tao
- Philip S. Yu
- Qingsong Wen
- Yang Liu