All Projects → hendrycks → Robustness

hendrycks / Robustness

Licence: apache-2.0
Corruption and Perturbation Robustness (ICLR 2019)

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to Robustness

Computer Vision
Programming Assignments and Lectures for Stanford's CS 231: Convolutional Neural Networks for Visual Recognition
Stars: ✭ 408 (-11.88%)
Mutual labels:  convolutional-neural-networks, imagenet
Models Comparison.pytorch
Code for the paper Benchmark Analysis of Representative Deep Neural Network Architectures
Stars: ✭ 148 (-68.03%)
Mutual labels:  convolutional-neural-networks, imagenet
Chainer Pspnet
PSPNet in Chainer
Stars: ✭ 76 (-83.59%)
Mutual labels:  convolutional-neural-networks, imagenet
Constrained attention filter
(ECCV 2020) Tensorflow implementation of A Generic Visualization Approach for Convolutional Neural Networks
Stars: ✭ 36 (-92.22%)
Mutual labels:  convolutional-neural-networks, imagenet
Pyconv
Pyramidal Convolution: Rethinking Convolutional Neural Networks for Visual Recognition (https://arxiv.org/pdf/2006.11538.pdf)
Stars: ✭ 231 (-50.11%)
Mutual labels:  convolutional-neural-networks, imagenet
Big transfer
Official repository for the "Big Transfer (BiT): General Visual Representation Learning" paper.
Stars: ✭ 1,096 (+136.72%)
Mutual labels:  convolutional-neural-networks, imagenet
Ghostnet
CV backbones including GhostNet, TinyNet and TNT, developed by Huawei Noah's Ark Lab.
Stars: ✭ 1,744 (+276.67%)
Mutual labels:  convolutional-neural-networks, imagenet
Tf Mobilenet V2
Mobilenet V2(Inverted Residual) Implementation & Trained Weights Using Tensorflow
Stars: ✭ 85 (-81.64%)
Mutual labels:  convolutional-neural-networks, imagenet
Triplet Attention
Official PyTorch Implementation for "Rotate to Attend: Convolutional Triplet Attention Module." [WACV 2021]
Stars: ✭ 222 (-52.05%)
Mutual labels:  convolutional-neural-networks, imagenet
Iresnet
Improved Residual Networks (https://arxiv.org/pdf/2004.04989.pdf)
Stars: ✭ 163 (-64.79%)
Mutual labels:  convolutional-neural-networks, imagenet
Switchable Normalization
Code for Switchable Normalization from "Differentiable Learning-to-Normalize via Switchable Normalization", https://arxiv.org/abs/1806.10779
Stars: ✭ 804 (+73.65%)
Mutual labels:  convolutional-neural-networks, imagenet
Pytorch Randaugment
Unofficial PyTorch Reimplementation of RandAugment.
Stars: ✭ 323 (-30.24%)
Mutual labels:  convolutional-neural-networks, imagenet
Addernet
Code for paper " AdderNet: Do We Really Need Multiplications in Deep Learning?"
Stars: ✭ 722 (+55.94%)
Mutual labels:  convolutional-neural-networks, imagenet
Pneumonia Detection From Chest X Ray Images With Deep Learning
Detecting Pneumonia in Chest X-ray Images using Convolutional Neural Network and Pretrained Models
Stars: ✭ 64 (-86.18%)
Mutual labels:  convolutional-neural-networks, imagenet
Caffenet Benchmark
Evaluation of the CNN design choices performance on ImageNet-2012.
Stars: ✭ 700 (+51.19%)
Mutual labels:  convolutional-neural-networks, imagenet
Imagenet
TensorFlow implementation of AlexNet and its training and testing on ImageNet ILSVRC 2012 dataset
Stars: ✭ 155 (-66.52%)
Mutual labels:  convolutional-neural-networks, imagenet
Assembled Cnn
Tensorflow implementation of "Compounding the Performance Improvements of Assembled Techniques in a Convolutional Neural Network"
Stars: ✭ 319 (-31.1%)
Mutual labels:  convolutional-neural-networks, imagenet
Espnetv2
A light-weight, power efficient, and general purpose convolutional neural network
Stars: ✭ 377 (-18.57%)
Mutual labels:  convolutional-neural-networks, imagenet
Multi Class Text Classification Cnn
Classify Kaggle Consumer Finance Complaints into 11 classes. Build the model with CNN (Convolutional Neural Network) and Word Embeddings on Tensorflow.
Stars: ✭ 410 (-11.45%)
Mutual labels:  convolutional-neural-networks
O Cnn
O-CNN: Octree-based Convolutional Neural Networks for 3D Shape Analysis
Stars: ✭ 432 (-6.7%)
Mutual labels:  convolutional-neural-networks

Benchmarking Neural Network Robustness to Common Corruptions and Perturbations

This repository contains the datasets and some code for the paper Benchmarking Neural Network Robustness to Common Corruptions and Perturbations (ICLR 2019) by Dan Hendrycks and Thomas Dietterich.

Requires Python 3+ and PyTorch 0.3+. For evaluation, please download the data from the links below.

ImageNet-C

Download ImageNet-C here. (Mirror.)

Download Tiny ImageNet-C here. (Mirror.)

Tiny ImageNet-C has 200 classes with images of size 64x64, while ImageNet-C has all 1000 classes where each image is the standard size. For even quicker experimentation, there is CIFAR-10-C and CIFAR-100-C. Evaluation using the JPEGs above is strongly prefered to computing the corruptions in memory, so that evaluation is deterministic and consistent.

ImageNet-C Leaderboard

ImageNet-C Robustness with a ResNet-50 Backbone trained on ImageNet-1K and evaluated on 224x224x3 images.

Method Reference Standalone? mCE Clean Error
DeepAugment+AugMix Hendrycks et al. No 53.6% 24.2%
Assemble-ResNet50 Lee et al. No 56.5% 17.90%
ANT (3x3) Rusak and Schott et al. Yes 63% 23.9%
BlurAfterConv Vasconcelos et al. Yes 64.9% 21.2%
AugMix Hendrycks and Mu et al. (ICLR 2020) Yes 65.3% 22.47%
Stylized ImageNet Geirhos et al. (ICLR 2019) Yes 69.3% 25.41%
Patch Uniform Lopes et al. Yes 74.3% 24.5%
ResNet-50 Baseline N/A 76.7% 23.85%

"Standalone" indicates whether the method is a combination of techniques or a standalone/single method. Combining methods and proposing standalone methods are both valuable but not necessarily commensurable.

Be sure to check each paper for results on all 15 corruptions, as some of these techniques improve robustness on all corruptions, some methods help on some corruptions and hurt on others, and some are exceptional against noise corruptions. Other backbones can obtain better results. For example, a vanilla ResNeXt-101 has an mCE of 62.2%. Note Lopes et al. have a ResNet-50 backbone with an mCE of 80.6, so their improvement is larger than what is immediately suggested by the table.

Submit a pull request if you beat the state-of-the-art on ImageNet-C with a ResNet-50 backbone.

UPDATE: New Robustness Benchmarks

For other distribution shift benchmarks like ImageNet-C, consider datasets like ImageNet-A or ImageNet-R.

ImageNet-A contains real-world, unmodified natural images that cause model accuracy to substantially degrade. ImageNet-R(endition) has 30,000 renditions of ImageNet classes cocering art, cartoons, deviantart, graffiti, embroidery, graphics, origami, paintings, patterns, plastic objects, plush objects, sculptures, sketches, tattoos, toys, and video games.

Calculating the mCE

This spreadsheet shows how to calculate the mean Corruption Error.

ImageNet-P

ImageNet-P sequences are MP4s not GIFs. The spatter perturbation sequence is a validation sequence.

Download Tiny ImageNet-P here. (Mirror.)

Download ImageNet-P here. (Mirror.)

ImageNet-P Leaderboard

ImageNet-P Perturbation Robustness with a ResNet-50 Backbone

Method Reference mFR mT5D
AugMix Hendrycks and Mu et al. (ICLR 2020) 37.4%
Low Pass Filter Pooling (bin-5) Zhang (ICML 2019) 51.2% 71.9%
ResNet-50 Baseline 58.0% 78.4%

Submit a pull request if you beat the state-of-the-art on ImageNet-P.

Citation

If you find this useful in your research, please consider citing:

@article{hendrycks2019robustness,
  title={Benchmarking Neural Network Robustness to Common Corruptions and Perturbations},
  author={Dan Hendrycks and Thomas Dietterich},
  journal={Proceedings of the International Conference on Learning Representations},
  year={2019}
}

Part of the code was contributed by Tom Brown.

Icons-50 (From an Older Draft)

Download Icons-50 here or here.

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].