All Projects → fwang91 → Residual Attention Network

fwang91 / Residual Attention Network

Residual Attention Network for Image Classification

Projects that are alternatives of or similar to Residual Attention Network

Global Self Attention Network
A Pytorch implementation of Global Self-Attention Network, a fully-attention backbone for vision tasks
Stars: ✭ 64 (-87.81%)
Mutual labels:  image-classification, attention
Wildcat.pytorch
PyTorch implementation of "WILDCAT: Weakly Supervised Learning of Deep ConvNets for Image Classification, Pointwise Localization and Segmentation", CVPR 2017
Stars: ✭ 238 (-54.67%)
Mutual labels:  image-classification, cvpr-2017
image-recognition
采用深度学习方法进行刀具识别。
Stars: ✭ 19 (-96.38%)
Mutual labels:  image-classification, attention
Pytorch Original Transformer
My implementation of the original transformer model (Vaswani et al.). I've additionally included the playground.py file for visualizing otherwise seemingly hard concepts. Currently included IWSLT pretrained models.
Stars: ✭ 411 (-21.71%)
Mutual labels:  attention
Gansformer
Generative Adversarial Transformers
Stars: ✭ 421 (-19.81%)
Mutual labels:  attention
Pba
Efficient Learning of Augmentation Policy Schedules
Stars: ✭ 461 (-12.19%)
Mutual labels:  image-classification
Gluon Cv
Gluon CV Toolkit
Stars: ✭ 5,001 (+852.57%)
Mutual labels:  image-classification
Metaoptnet
Meta-Learning with Differentiable Convex Optimization (CVPR 2019 Oral)
Stars: ✭ 412 (-21.52%)
Mutual labels:  image-classification
Punctuator2
A bidirectional recurrent neural network model with attention mechanism for restoring missing punctuation in unsegmented text
Stars: ✭ 483 (-8%)
Mutual labels:  attention
Caer
High-performance Vision library in Python. Scale your research, not boilerplate.
Stars: ✭ 452 (-13.9%)
Mutual labels:  image-classification
Ban Vqa
Bilinear attention networks for visual question answering
Stars: ✭ 449 (-14.48%)
Mutual labels:  attention
Tfclassifier
Tensorflow based training and classification scripts for text, images, etc
Stars: ✭ 441 (-16%)
Mutual labels:  image-classification
Rnn Nlu
A TensorFlow implementation of Recurrent Neural Networks for Sequence Classification and Sequence Labeling
Stars: ✭ 463 (-11.81%)
Mutual labels:  attention
Recurrent Visual Attention
A PyTorch Implementation of "Recurrent Models of Visual Attention"
Stars: ✭ 414 (-21.14%)
Mutual labels:  attention
Bag of tricks for image classification with convolutional neural networks
experiments on Paper <Bag of Tricks for Image Classification with Convolutional Neural Networks> and other useful tricks to improve CNN acc
Stars: ✭ 486 (-7.43%)
Mutual labels:  image-classification
Neural Backed Decision Trees
Making decision trees competitive with neural networks on CIFAR10, CIFAR100, TinyImagenet200, Imagenet
Stars: ✭ 411 (-21.71%)
Mutual labels:  image-classification
Deepalignmentnetwork
A deep neural network for face alignment
Stars: ✭ 480 (-8.57%)
Mutual labels:  cvpr-2017
Openpose
OpenPose: Real-time multi-person keypoint detection library for body, face, hands, and foot estimation
Stars: ✭ 22,892 (+4260.38%)
Mutual labels:  cvpr-2017
Seefood
Inspired by HBO's Silicon Valley: SeeFood is an iOS app that uses CoreML to detect various dishes
Stars: ✭ 445 (-15.24%)
Mutual labels:  image-classification
Structured Self Attention
A Structured Self-attentive Sentence Embedding
Stars: ✭ 459 (-12.57%)
Mutual labels:  attention

Residual Attention Network

Residual Attention Network for Image Classification (CVPR-2017 Spotlight)

By Fei Wang, Mengqing Jiang, Chen Qian, Shuo Yang, Chen Li, Honggang Zhang, Xiaogang Wang, Xiaoou Tang

Introduction

Residual Attention Network is a convolutional neural network using attention mechanism which can incorporate with state-of-the-art feed forward network architecture in an end-to-end training fashion.

Residual Attention Networks are described in the paper "Residual Attention Network for Image Classification"(https://arxiv.org/pdf/1704.06904.pdf).

This repository contains the prototxts of "Residual Attention Network". The trained model will be released soon.

Citation

If you find "Residual Attention Network" useful in your research, please cite:

@article{wang2017residual,
	title={Residual Attention Network for Image Classification},
	author={Wang, Fei and Jiang, Mengqing and Qian, Chen and Yang, Shuo and Li, Cheng and Zhang, Honggang and Wang, Xiaogang and Tang, Xiaoou},
	journal={arXiv preprint arXiv:1704.06904},
	year={2017}	
}

Models

  1. Attention-56 and Attention-92 are based on the pre-activation residual unit.

  2. According to the paper, we replace pre-activation residual unit with resnext unit to contruct the AttentionNeXt-56 and AttentionNeXt-92.

Main Performance

  1. Evaluation on ImageNet validation dataset.
Network Test Size top-1 top-5
Attention-56 224*224 21.76% 5.9%
AttentionNeXt-56 224*224 21.2% 5.6%
Attention-92 320*320 19.5% 4.8%

Note

  1. We use Caffe ("https://github.com/buptwangfei/caffe") to train our Residual Attention Networks.

  2. We follow the implementation of BN layer from "https://github.com/happynear/caffe-windows.git", which merge computations of mean, variance, scale and shift into one layer. We use moving average in the training stage.

  3. The scale augmentation and ratio augmentation are used in the training process.

  4. The mini-batch of per GPU should be at least 32 images.

  5. If you want to train Residual Attention Network, you should use my caffe code and add data augmentation described in the paper. I think it is easy to reproduce the performance on the ImageNet validation dataset.

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].