All Projects → gaohuang → Snapshotensemble

gaohuang / Snapshotensemble

Licence: bsd-3-clause
Snapshot Ensembles in Torch (Snapshot Ensembles: Train 1, Get M for Free)

Programming Languages

lua
6591 projects

Projects that are alternatives of or similar to Snapshotensemble

Tf texture net
TensorFlow implementation of DmitryUlyanov/texture_nets
Stars: ✭ 96 (-44.51%)
Mutual labels:  torch
Cyclegan
Software that can generate photos from paintings, turn horses into zebras, perform style transfer, and more.
Stars: ✭ 10,933 (+6219.65%)
Mutual labels:  torch
Samplernn torch
Torch implementation of SampleRNN: An Unconditional End-to-End Neural Audio Generation Model
Stars: ✭ 146 (-15.61%)
Mutual labels:  torch
Pytorch Learners Tutorial
PyTorch tutorial for learners
Stars: ✭ 97 (-43.93%)
Mutual labels:  torch
Deepnudecli
DeepNude Command Line Version With Watermark Removed
Stars: ✭ 112 (-35.26%)
Mutual labels:  torch
Neural Style Audio Torch
Torch implementation for audio neural style.
Stars: ✭ 130 (-24.86%)
Mutual labels:  torch
Beauty.torch
Understanding facial beauty with deep learning.
Stars: ✭ 90 (-47.98%)
Mutual labels:  torch
Semantic3dnet
Point cloud semantic segmentation via Deep 3D Convolutional Neural Network
Stars: ✭ 170 (-1.73%)
Mutual labels:  torch
Ti Pooling
TI-pooling: transformation-invariant pooling for feature learning in Convolutional Neural Networks
Stars: ✭ 119 (-31.21%)
Mutual labels:  torch
Synthesize3dviadepthorsil
[CVPR 2017] Generation and reconstruction of 3D shapes via modeling multi-view depth maps or silhouettes
Stars: ✭ 141 (-18.5%)
Mutual labels:  torch
Torchelie
Torchélie is a set of utility functions, layers, losses, models, trainers and other things for PyTorch.
Stars: ✭ 98 (-43.35%)
Mutual labels:  torch
Laia
Laia: A deep learning toolkit for HTR based on Torch
Stars: ✭ 109 (-36.99%)
Mutual labels:  torch
Torchsample
High-Level Training, Data Augmentation, and Utilities for Pytorch
Stars: ✭ 1,731 (+900.58%)
Mutual labels:  torch
Torchcraft
Connecting Torch to StarCraft
Stars: ✭ 1,341 (+675.14%)
Mutual labels:  torch
Skip Thoughts.torch
Porting of Skip-Thoughts pretrained models from Theano to PyTorch & Torch7
Stars: ✭ 146 (-15.61%)
Mutual labels:  torch
Grad Cam
🌈 📷 Gradient-weighted Class Activation Mapping (Grad-CAM) Demo
Stars: ✭ 91 (-47.4%)
Mutual labels:  torch
Ganspapercollection
Stars: ✭ 130 (-24.86%)
Mutual labels:  torch
Torch Dct
DCT (discrete cosine transform) functions for pytorch
Stars: ✭ 173 (+0%)
Mutual labels:  torch
Dockerfiles
Deep Learning Dockerfiles
Stars: ✭ 150 (-13.29%)
Mutual labels:  torch
Prediction Flow
Deep-Learning based CTR models implemented by PyTorch
Stars: ✭ 138 (-20.23%)
Mutual labels:  torch

Snapshot Ensembles: Train 1, Get M for Free

This repository contains the Torch code for the paper Snapshot Ensembles: Train 1, Get M for Free.

The code is based on fb.resnet.torch by Facebook .

There is also a nice Keras implementation by titu1994.

Table of Contents

  1. Introduction
  2. Usage
  3. Contact

Introduction

Snapshot Ensemble is a method to obtain ensembles of multiple neural network at no additional training cost. This is achieved by letting a single neural network converge into several local minima along its optimization path and save the model parameters. The repeated rapid convergence is realized using multiple learning rate annealing cycles.

Figure 1: Left: Illustration of SGD optimization with a typical learning rate schedule. The model converges to a minimum at the end of training. Right: Illustration of Snapshot Ensembling optimization. The model undergoes several learning rate annealing cycles, converging to and escaping from multiple local minima. We take a snapshot at each minimum for test time ensembling.

Usage

  1. Install Torch ResNet (https://github.com/facebook/fb.resnet.torch);
  2. Clone the files to the fb.resnet.torch/ directory. Note that you need to replace train.lua with the one from this repository;
  3. An example command to train a Snapshot Ensemble with ResNet-110 (B = 200 epochs, M = 5 cycles, Initial learning rate alpha = 0.2) on CIFAR-100:

th main.lua -netType resnet -depth 110 -dataset cifar100 -batchSize 64 -nEpochs 200 -lrShape cosine -nCycles 5 -LR 0.2 -save checkpoints/

Contact

[gh349, yl2363] at cornell.edu   Any discussions, suggestions and questions are welcome!

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].