Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Diagnostics for Deep Neural Networks with Automated Copy/Paste Attacks
1
Zitationen
3
Autoren
2022
Jahr
Abstract
This paper considers the problem of helping humans exercise scalable oversight over deep neural networks (DNNs). Adversarial examples can be useful by helping to reveal weaknesses in DNNs, but they can be difficult to interpret or draw actionable conclusions from. Some previous works have proposed using human-interpretable adversarial attacks including copy/paste attacks in which one natural image pasted into another causes an unexpected misclassification. We build on these with two contributions. First, we introduce Search for Natural Adversarial Features Using Embeddings (SNAFUE) which offers a fully automated method for finding copy/paste attacks. Second, we use SNAFUE to red team an ImageNet classifier. We reproduce copy/paste attacks from previous works and find hundreds of other easily-describable vulnerabilities, all without a human in the loop. Code is available at https://github.com/thestephencasper/snafue
Ähnliche Arbeiten
Rethinking the Inception Architecture for Computer Vision
2016 · 30.409 Zit.
MobileNetV2: Inverted Residuals and Linear Bottlenecks
2018 · 24.540 Zit.
CBAM: Convolutional Block Attention Module
2018 · 21.437 Zit.
An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale
2020 · 21.345 Zit.
Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification
2015 · 18.533 Zit.