Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Evaluating Large Language Models Security and Resilience: A Practical Testing Framework
0
Zitationen
3
Autoren
2025
Jahr
Abstract
Large Language Models (LLMs) are increasingly used in real-world applications, but as their capabilities grow, so do the risks of misuse. Despite their widespread adoption, the security of these models remains an area with many open questions. This paper explores these issues through a set of applied experiments carried out in a controlled environment designed for testing. A prototype application that allows demonstrating how an LLM security benchmarking tool could function in practice was designed. The application allows users to simulate attacks and assess the effectiveness of several defense strategies as in-context defense and paraphrase-based. The experimental results show notable differences between the tested methods. Some techniques were able to fully block attacks while maintaining the model’ ability to respond accurately to regular prompts. Our work paves the way for a more secure development of LLMs by evaluating their resilience to known attacks, while also providing a practical prototype that serves as a starting point for future research and can be extended to support more advanced evaluation methodologies in the context of security of generative AI systems.
Ähnliche Arbeiten
Rethinking the Inception Architecture for Computer Vision
2016 · 30.366 Zit.
MobileNetV2: Inverted Residuals and Linear Bottlenecks
2018 · 24.462 Zit.
CBAM: Convolutional Block Attention Module
2018 · 21.360 Zit.
An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale
2020 · 21.317 Zit.
Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification
2015 · 18.510 Zit.