All Projects → hongyi-zhang → Fixup

hongyi-zhang / Fixup

Licence: bsd-3-clause
A Re-implementation of Fixed-update Initialization

Programming Languages

python
139335 projects - #7 most used programming language

Labels

Projects that are alternatives of or similar to Fixup

Tools To Design Or Visualize Architecture Of Neural Network
Tools to Design or Visualize Architecture of Neural Network
Stars: ✭ 1,143 (+836.89%)
Mutual labels:  resnet
Hypernetworks
PyTorch implementation of HyperNetworks (Ha et al., ICLR 2017) for ResNet (Residual Networks)
Stars: ✭ 95 (-22.13%)
Mutual labels:  resnet
Numpy Ml
Machine learning, in numpy
Stars: ✭ 11,100 (+8998.36%)
Mutual labels:  resnet
Yolo resnet
Implementing YOLO using ResNet as the feature extraction network
Stars: ✭ 82 (-32.79%)
Mutual labels:  resnet
Pytorch Classification
Classification with PyTorch.
Stars: ✭ 1,268 (+939.34%)
Mutual labels:  resnet
Deep Reinforcement Learning With Pytorch
PyTorch implementation of DQN, AC, ACER, A2C, A3C, PG, DDPG, TRPO, PPO, SAC, TD3 and ....
Stars: ✭ 1,345 (+1002.46%)
Mutual labels:  resnet
Deep Ranking
Learning Fine-grained Image Similarity with Deep Ranking is a novel application of neural networks, where the authors use a new multi scale architecture combined with a triplet loss to create a neural network that is able to perform image search. This repository is a simplified implementation of the same
Stars: ✭ 64 (-47.54%)
Mutual labels:  resnet
Ir Net
This project is the PyTorch implementation of our accepted CVPR 2020 paper : forward and backward information retention for accurate binary neural networks.
Stars: ✭ 119 (-2.46%)
Mutual labels:  resnet
Keras Resnet
Residual networks implementation using Keras-1.0 functional API
Stars: ✭ 1,286 (+954.1%)
Mutual labels:  resnet
Resnet Face Pytorch
Training a ResNet on UMDFaces for face recognition
Stars: ✭ 107 (-12.3%)
Mutual labels:  resnet
Bsconv
Reference implementation for Blueprint Separable Convolutions (CVPR 2020)
Stars: ✭ 84 (-31.15%)
Mutual labels:  resnet
M Pact
A one stop shop for all of your activity recognition needs.
Stars: ✭ 85 (-30.33%)
Mutual labels:  resnet
Tensorflow2.0 Examples
🙄 Difficult algorithm, Simple code.
Stars: ✭ 1,397 (+1045.08%)
Mutual labels:  resnet
Boring Detector
State-of-the-art detector of Boring hats in images and videos.
Stars: ✭ 79 (-35.25%)
Mutual labels:  resnet
Shot Type Classifier
Detecting cinema shot types using a ResNet-50
Stars: ✭ 109 (-10.66%)
Mutual labels:  resnet
Gluon2pytorch
Gluon to PyTorch deep neural network model converter
Stars: ✭ 70 (-42.62%)
Mutual labels:  resnet
Holocron
PyTorch implementations of recent Computer Vision tricks
Stars: ✭ 95 (-22.13%)
Mutual labels:  resnet
Pyramidnet
Torch implementation of the paper "Deep Pyramidal Residual Networks" (https://arxiv.org/abs/1610.02915).
Stars: ✭ 121 (-0.82%)
Mutual labels:  resnet
Shiftresnet Cifar
ResNet with Shift, Depthwise, or Convolutional Operations for CIFAR-100, CIFAR-10 on PyTorch
Stars: ✭ 112 (-8.2%)
Mutual labels:  resnet
Resnet Imagenet Caffe
train resnet on imagenet from scratch with caffe
Stars: ✭ 105 (-13.93%)
Mutual labels:  resnet

Fixup

A Re-implementation of Fixed-update Initialization (https://arxiv.org/abs/1901.09321). (requires Pytorch 1.0)

Cite as:

Hongyi Zhang, Yann N. Dauphin, Tengyu Ma. Fixup Initialization: Residual Learning Without Normalization. 7th International Conference on Learning Representations (ICLR 2019).


ResNet for CIFAR-10

The default arguments will train a ResNet-110 (https://arxiv.org/abs/1512.03385) with Fixup + Mixup (https://arxiv.org/abs/1710.09412).

Example:

The following script will train a ResNet-32 model (https://arxiv.org/abs/1512.03385) on GPU 0 with Fixup and no Mixup (alpha=0), with weight decay 5e-4 and (the default) learning rate 0.1 and batch size 128.

CUDA_VISIBLE_DEVICES=0 python cifar_train.py -a fixup_resnet32 --sess benchmark_a0d5e4lr01 --seed 11111 --alpha 0. --decay 5e-4

ResNet for ImageNet

ImageNet models with training scripts are now available. (Thanks @tjingrant for help!)

Top-1 accuracy for ResNet-50 at Epoch 100 with Mixup (alpha=0.7) is around 76.0%.


Transformer for machine translation

Transformer model with Fixup (instead of layer normalization) is available. To run the experiments, you will need to download and install the fairseq library (the provided code was tested on an earlier version: https://github.com/pytorch/fairseq/tree/5d00e8eea2644611f397d05c6c8f15083388b8b4). You can then copy the files into corresponding folders.

An example script run.sh is provided to run the IWSLT experiments described in the paper. For more information, please refer to the instructions in fairseq repo (https://github.com/pytorch/fairseq/tree/5d00e8eea2644611f397d05c6c8f15083388b8b4/examples/translation).

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].