Detecting AutoAttack Perturbations in the Frequency Domain
- URL: http://arxiv.org/abs/2111.08785v3
- Date: Tue, 20 Feb 2024 13:42:11 GMT
- Title: Detecting AutoAttack Perturbations in the Frequency Domain
- Authors: Peter Lorenz, Paula Harder, Dominik Strassel, Margret Keuper and Janis
Keuper
- Abstract summary: adversarial attacks on image classification networks by the AutoAttack framework have drawn a lot of attention.
In this paper, we investigate the spatial and frequency domain properties of AutoAttack and propose an alternative defense.
Instead of hardening a network, we detect adversarial attacks during inference, rejecting manipulated inputs.
- Score: 18.91242463856906
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recently, adversarial attacks on image classification networks by the
AutoAttack (Croce and Hein, 2020b) framework have drawn a lot of attention.
While AutoAttack has shown a very high attack success rate, most defense
approaches are focusing on network hardening and robustness enhancements, like
adversarial training. This way, the currently best-reported method can
withstand about 66% of adversarial examples on CIFAR10. In this paper, we
investigate the spatial and frequency domain properties of AutoAttack and
propose an alternative defense. Instead of hardening a network, we detect
adversarial attacks during inference, rejecting manipulated inputs. Based on a
rather simple and fast analysis in the frequency domain, we introduce two
different detection algorithms. First, a black box detector that only operates
on the input images and achieves a detection accuracy of 100% on the AutoAttack
CIFAR10 benchmark and 99.3% on ImageNet, for epsilon = 8/255 in both cases.
Second, a whitebox detector using an analysis of CNN feature maps, leading to a
detection rate of also 100% and 98.7% on the same benchmarks.
Related papers
- AdvQDet: Detecting Query-Based Adversarial Attacks with Adversarial Contrastive Prompt Tuning [93.77763753231338]
Adversarial Contrastive Prompt Tuning (ACPT) is proposed to fine-tune the CLIP image encoder to extract similar embeddings for any two intermediate adversarial queries.
We show that ACPT can detect 7 state-of-the-art query-based attacks with $>99%$ detection rate within 5 shots.
We also show that ACPT is robust to 3 types of adaptive attacks.
arXiv Detail & Related papers (2024-08-04T09:53:50Z) - New Adversarial Image Detection Based on Sentiment Analysis [37.139957973240264]
adversarial attack models, e.g., DeepFool, are on the rise and outrunning adversarial example detection techniques.
This paper presents a new adversarial example detector that outperforms state-of-the-art detectors in identifying the latest adversarial attacks on image datasets.
arXiv Detail & Related papers (2023-05-03T14:32:21Z) - Is RobustBench/AutoAttack a suitable Benchmark for Adversarial
Robustness? [20.660465258314314]
We argue that the alternation of data by AutoAttack with l-inf, eps = 8/255 is unrealistically strong, resulting in close to perfect detection rates of adversarial samples.
We also show that other attack methods are much harder to detect while achieving similar success rates.
arXiv Detail & Related papers (2021-12-02T20:44:16Z) - SpectralDefense: Detecting Adversarial Attacks on CNNs in the Fourier
Domain [10.418647759223964]
We show how analysis in the Fourier domain of input images and feature maps can be used to distinguish benign test samples from adversarial images.
We propose two novel detection methods.
arXiv Detail & Related papers (2021-03-04T12:48:28Z) - Detection of Adversarial Supports in Few-shot Classifiers Using Feature
Preserving Autoencoders and Self-Similarity [89.26308254637702]
We propose a detection strategy to highlight adversarial support sets.
We make use of feature preserving autoencoder filtering and also the concept of self-similarity of a support set to perform this detection.
Our method is attack-agnostic and also the first to explore detection for few-shot classifiers to the best of our knowledge.
arXiv Detail & Related papers (2020-12-09T14:13:41Z) - Anchor-free Small-scale Multispectral Pedestrian Detection [88.7497134369344]
We propose a method for effective and efficient multispectral fusion of the two modalities in an adapted single-stage anchor-free base architecture.
We aim at learning pedestrian representations based on object center and scale rather than direct bounding box predictions.
Results show our method's effectiveness in detecting small-scaled pedestrians.
arXiv Detail & Related papers (2020-08-19T13:13:01Z) - Bayesian Optimization with Machine Learning Algorithms Towards Anomaly
Detection [66.05992706105224]
In this paper, an effective anomaly detection framework is proposed utilizing Bayesian Optimization technique.
The performance of the considered algorithms is evaluated using the ISCX 2012 dataset.
Experimental results show the effectiveness of the proposed framework in term of accuracy rate, precision, low-false alarm rate, and recall.
arXiv Detail & Related papers (2020-08-05T19:29:35Z) - Anomaly Detection-Based Unknown Face Presentation Attack Detection [74.4918294453537]
Anomaly detection-based spoof attack detection is a recent development in face Presentation Attack Detection.
In this paper, we present a deep-learning solution for anomaly detection-based spoof attack detection.
The proposed approach benefits from the representation learning power of the CNNs and learns better features for fPAD task.
arXiv Detail & Related papers (2020-07-11T21:20:55Z) - Non-Intrusive Detection of Adversarial Deep Learning Attacks via
Observer Networks [5.4572790062292125]
Recent studies have shown that deep learning models are vulnerable to crafted adversarial inputs.
We propose a novel method to detect adversarial inputs by augmenting the main classification network with multiple binary detectors.
We achieve a 99.5% detection accuracy on the MNIST dataset and 97.5% on the CIFAR-10 dataset.
arXiv Detail & Related papers (2020-02-22T21:13:00Z) - R-FCN: Object Detection via Region-based Fully Convolutional Networks [87.62557357527861]
We present region-based, fully convolutional networks for accurate and efficient object detection.
Our result is achieved at a test-time speed of 170ms per image, 2.5-20x faster than the Faster R-CNN counterpart.
arXiv Detail & Related papers (2016-05-20T15:50:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.