2-Detection

This category of tools aim for improving the robustness of classifiers by developing general techniques detecting adversarial attacks..
This includes:


Feature Squeezing Mitigates and Detects Carlini-Wagner Adversarial Examples

Paper Arxiv

Abstract

Feature squeezing is a recently-introduced framework for mitigating and detecting adversarial examples. In previous work, we showed that it is effective against several earlier methods for generating adversarial examples. In this short note, we report on recent results showing that simple feature squeezing techniques also make deep learning models significantly more robust against the Carlini/Wagner attacks, which are the best known adversarial methods discovered to date.

Citations

@article{xu2017feature,
  title={Feature Squeezing Mitigates and Detects Carlini/Wagner Adversarial Examples},
  author={Xu, Weilin and Evans, David and Qi, Yanjun},
  journal={arXiv preprint arXiv:1705.10686},
  year={2017}
}

Support or Contact

Having troubl with our tools? Please contact Weilin and we’ll help you sort it out.


Feature Squeezing- Detecting Adversarial Examples in Deep Neural Networks

GitHub: FeatureSqueezing

Paper Arxiv

Abstract

Although deep neural networks (DNNs) have achieved great success in many computer vision tasks, recent studies have shown they are vulnerable to adversarial examples. Such examples, typically generated by adding small but purposeful distortions, can frequently fool DNN models. Previous studies to defend against adversarial examples mostly focused on refining the DNN models. They have either shown limited success or suffer from the expensive computation. We propose a new strategy, \emph{feature squeezing}, that can be used to harden DNN models by detecting adversarial examples. Feature squeezing reduces the search space available to an adversary by coalescing samples that correspond to many different feature vectors in the original space into a single sample. By comparing a DNN model’s prediction on the original input with that on the squeezed input, feature squeezing detects adversarial examples with high accuracy and few false positives. This paper explores two instances of feature squeezing: reducing the color bit depth of each pixel and smoothing using a spatial filter. These strategies are straightforward, inexpensive, and complementary to defensive methods that operate on the underlying model, such as adversarial training.

evadePDF

Citations

@article{xu2017feature,
  title={Feature Squeezing: Detecting Adversarial Examples in Deep Neural Networks},
  author={Xu, Weilin and Evans, David and Qi, Yanjun},
  journal={arXiv preprint arXiv:1704.01155},
  year={2017}
}

Support or Contact

Having troubl with our tools? Please contact Weilin and we’ll help you sort it out.