Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
The Limitations of Deep Learning in Adversarial Settings
3.853
Zitationen
6
Autoren
2016
Jahr
Abstract
Deep learning takes advantage of large datasets and computationally efficient training algorithms to outperform other approaches at various machine learning tasks. However, imperfections in the training phase of deep neural networks make them vulnerable to adversarial samples: inputs crafted by adversaries with the intent of causing deep neural networks to misclassify. In this work, we formalize the space of adversaries against deep neural networks (DNNs) and introduce a novel class of algorithms to craft adversarial samples based on a precise understanding of the mapping between inputs and outputs of DNNs. In an application to computer vision, we show that our algorithms can reliably produce samples correctly classified by human subjects but misclassified in specific targets by a DNN with a 97% adversarial success rate while only modifying on average 4.02% of the input features per sample. We then evaluate the vulnerability of different sample classes to adversarial perturbations by defining a hardness measure. Finally, we describe preliminary work outlining defenses against adversarial samples by defining a predictive measure of distance between a benign input and a target classification.
Ähnliche Arbeiten
Rethinking the Inception Architecture for Computer Vision
2016 · 30.290 Zit.
MobileNetV2: Inverted Residuals and Linear Bottlenecks
2018 · 24.337 Zit.
An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale
2020 · 21.280 Zit.
CBAM: Convolutional Block Attention Module
2018 · 21.197 Zit.
Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification
2015 · 18.473 Zit.