All Projects → anishathalye → Obfuscated Gradients

anishathalye / Obfuscated Gradients

Obfuscated Gradients Give a False Sense of Security: Circumventing Defenses to Adversarial Examples

Projects that are alternatives of or similar to Obfuscated Gradients

Ml Course Msu
Lecture notes and code for Machine Learning practical course on CMC MSU
Stars: ✭ 759 (-2.69%)
Mutual labels:  jupyter-notebook
Pytorch 3dunet
3D U-Net model for volumetric semantic segmentation written in pytorch
Stars: ✭ 765 (-1.92%)
Mutual labels:  jupyter-notebook
Clstm
A small C++ implementation of LSTM networks, focused on OCR.
Stars: ✭ 771 (-1.15%)
Mutual labels:  jupyter-notebook
Tffm
TensorFlow implementation of an arbitrary order Factorization Machine
Stars: ✭ 761 (-2.44%)
Mutual labels:  jupyter-notebook
996quant
35岁程序员退路之量化投资学习笔记
Stars: ✭ 766 (-1.79%)
Mutual labels:  jupyter-notebook
Boltzmann Machines
Boltzmann Machines in TensorFlow with examples
Stars: ✭ 768 (-1.54%)
Mutual labels:  jupyter-notebook
Entity Embedding Rossmann
Stars: ✭ 758 (-2.82%)
Mutual labels:  jupyter-notebook
Python Machine Learning Book 2nd Edition
The "Python Machine Learning (2nd edition)" book code repository and info resource
Stars: ✭ 6,422 (+723.33%)
Mutual labels:  jupyter-notebook
Opacus
Training PyTorch models with differential privacy
Stars: ✭ 763 (-2.18%)
Mutual labels:  jupyter-notebook
Estool
Evolution Strategies Tool
Stars: ✭ 767 (-1.67%)
Mutual labels:  jupyter-notebook
Jupyter2slides
Cloud Native Presentation Slides with Jupyter Notebook + Reveal.js
Stars: ✭ 762 (-2.31%)
Mutual labels:  jupyter-notebook
Ktrain
ktrain is a Python library that makes deep learning and AI more accessible and easier to apply
Stars: ✭ 763 (-2.18%)
Mutual labels:  jupyter-notebook
Coloring Greyscale Images
Coloring black and white images with deep learning.
Stars: ✭ 767 (-1.67%)
Mutual labels:  jupyter-notebook
Ec2 Spot Labs
Collection of tools and code examples to demonstrate best practices in using Amazon EC2 Spot Instances.
Stars: ✭ 758 (-2.82%)
Mutual labels:  jupyter-notebook
Coursera
Quiz & Assignment of Coursera
Stars: ✭ 774 (-0.77%)
Mutual labels:  jupyter-notebook
Notedown
Markdown <=> IPython Notebook
Stars: ✭ 757 (-2.95%)
Mutual labels:  jupyter-notebook
Deeplab v3
Tensorflow Implementation of the Semantic Segmentation DeepLab_V3 CNN
Stars: ✭ 768 (-1.54%)
Mutual labels:  jupyter-notebook
Deepmind Research
This repository contains implementations and illustrative code to accompany DeepMind publications
Stars: ✭ 9,190 (+1078.21%)
Mutual labels:  jupyter-notebook
Face Mask Detection
Face Mask Detection system based on computer vision and deep learning using OpenCV and Tensorflow/Keras
Stars: ✭ 774 (-0.77%)
Mutual labels:  jupyter-notebook
Hands On Meta Learning With Python
Learning to Learn using One-Shot Learning, MAML, Reptile, Meta-SGD and more with Tensorflow
Stars: ✭ 768 (-1.54%)
Mutual labels:  jupyter-notebook

Obfuscated Gradients Give a False Sense of Security: Circumventing Defenses to Adversarial Examples

Above is an adversarial example: the slightly perturbed image of the cat fools an InceptionV3 classifier into classifying it as "guacamole". Such "fooling images" are easy to synthesize using gradient descent (Szegedy et al. 2013).

In our recent paper, we evaluate the robustness of nine papers accepted to ICLR 2018 as non-certified white-box-secure defenses to adversarial examples. We find that seven of the nine defenses provide a limited increase in robustness and can be broken by improved attack techniques we develop.

Below is Table 1 from our paper, where we show the robustness of each accepted defense to the adversarial examples we can construct:

Defense Dataset Distance Accuracy
Buckman et al. (2018) CIFAR 0.031 (linf) 0%*
Ma et al. (2018) CIFAR 0.031 (linf) 5%
Guo et al. (2018) ImageNet 0.05 (l2) 0%*
Dhillon et al. (2018) CIFAR 0.031 (linf) 0%
Xie et al. (2018) ImageNet 0.031 (linf) 0%*
Song et al. (2018) CIFAR 0.031 (linf) 9%*
Samangouei et al. (2018) MNIST 0.005 (l2) 55%**
Madry et al. (2018) CIFAR 0.031 (linf) 47%
Na et al. (2018) CIFAR 0.015 (linf) 15%

(Defenses denoted with * also propose combining adversarial training; we report here the defense alone. See our paper, Section 5 for full numbers. The fundemental principle behind the defense denoted with ** has 0% accuracy; in practice defense imperfections cause the theoretically optimal attack to fail, see Section 5.4.2 for details.)

The only defense we observe that significantly increases robustness to adversarial examples within the threat model proposed is "Towards Deep Learning Models Resistant to Adversarial Attacks" (Madry et al. 2018), and we were unable to defeat this defense without stepping outside the threat model. Even then, this technique has been shown to be difficult to scale to ImageNet-scale (Kurakin et al. 2016). The remainder of the papers (besides the paper by Na et al., which provides limited robustness) rely either inadvertently or intentionally on what we call obfuscated gradients. Standard attacks apply gradient descent to maximize the loss of the network on a given image to generate an adversarial example on a neural network. Such optimization methods require a useful gradient signal to succeed. When a defense obfuscates gradients, it breaks this gradient signal and causes optimization based methods to fail.

We identify three ways in which defenses cause obfuscated gradients, and construct attacks to bypass each of these cases. Our attacks are generally applicable to any defense that includes, either intentionally or or unintentionally, a non-differentiable operation or otherwise prevents gradient signal from flowing through the network. We hope future work will be able to use our approaches to perform a more thorough security evaluation.

Paper

Abstract:

We identify obfuscated gradients, a kind of gradient masking, as a phenomenon that leads to a false sense of security in defenses against adversarial examples. While defenses that cause obfuscated gradients appear to defeat iterative optimization-based attacks, we find defenses relying on this effect can be circumvented. We describe characteristic behaviors of defenses exhibiting the effect, and for each of the three types of obfuscated gradients we discover, we develop attack techniques to overcome it. In a case study, examining non-certified white-box-secure defenses at ICLR 2018, we find obfuscated gradients are a common occurrence, with 7 of 9 defenses relying on obfuscated gradients. Our new attacks successfully circumvent 6 completely, and 1 partially, in the original threat model each paper considers.

For details, read our paper.

Source code

This repository contains our instantiations of the general attack techniques described in our paper, breaking 7 of the ICLR 2018 defenses. Some of the defenses didn't release source code (at the time we did this work), so we had to reimplement them.

Citation

@inproceedings{obfuscated-gradients,
  author = {Anish Athalye and Nicholas Carlini and David Wagner},
  title = {Obfuscated Gradients Give a False Sense of Security: Circumventing Defenses to Adversarial Examples},
  booktitle = {Proceedings of the 35th International Conference on Machine Learning, {ICML} 2018},
  year = {2018},
  month = jul,
  url = {https://arxiv.org/abs/1802.00420},
}
Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].