huyvnphan / Pytorch_cifar10
Licence: mit
Pretrained TorchVision models on CIFAR10 dataset (with weights)
Stars: ✭ 219
Programming Languages
python
139335 projects - #7 most used programming language
Projects that are alternatives of or similar to Pytorch cifar10
Relativistic Average Gan Keras
The implementation of Relativistic average GAN with Keras
Stars: ✭ 36 (-83.56%)
Mutual labels: deeplearning, cifar10
Neural Backed Decision Trees
Making decision trees competitive with neural networks on CIFAR10, CIFAR100, TinyImagenet200, Imagenet
Stars: ✭ 411 (+87.67%)
Mutual labels: pretrained-models, cifar10
Dawn Bench Entries
DAWNBench: An End-to-End Deep Learning Benchmark and Competition
Stars: ✭ 254 (+15.98%)
Mutual labels: deeplearning, cifar10
Spectralnormalizationkeras
Spectral Normalization for Keras Dense and Convolution Layers
Stars: ✭ 100 (-54.34%)
Mutual labels: deeplearning, cifar10
Netron
Visualizer for neural network, deep learning, and machine learning models
Stars: ✭ 17,193 (+7750.68%)
Mutual labels: deeplearning
Keras Unet
Helper package with multiple U-Net implementations in Keras as well as useful utility tools helpful when working with image semantic segmentation tasks. This library and underlying tools come from multiple projects I performed working on semantic segmentation tasks
Stars: ✭ 196 (-10.5%)
Mutual labels: deeplearning
Hyperspectral Classification
Hyperspectral-Classification Pytorch
Stars: ✭ 195 (-10.96%)
Mutual labels: deeplearning
Awesome Pretrained Chinese Nlp Models
Awesome Pretrained Chinese NLP Models,高质量中文预训练模型集合
Stars: ✭ 195 (-10.96%)
Mutual labels: pretrained-models
Cartoonize
A demo webapp to convert images and videos into cartoon!
Stars: ✭ 215 (-1.83%)
Mutual labels: deeplearning
Dialogrpt
EMNLP 2020: "Dialogue Response Ranking Training with Large-Scale Human Feedback Data"
Stars: ✭ 216 (-1.37%)
Mutual labels: pretrained-models
Awesome Deep Learning And Machine Learning Questions
【不定期更新】收集整理的一些网站中(如知乎、Quora、Reddit、Stack Exchange等)与深度学习、机器学习、强化学习、数据科学相关的有价值的问题
Stars: ✭ 203 (-7.31%)
Mutual labels: deeplearning
Sparse Structured Attention
Sparse and structured neural attention mechanisms
Stars: ✭ 198 (-9.59%)
Mutual labels: deeplearning
Clearml
ClearML - Auto-Magical CI/CD to streamline your ML workflow. Experiment Manager, MLOps and Data-Management
Stars: ✭ 2,868 (+1209.59%)
Mutual labels: deeplearning
Text Classification
Text Classification through CNN, RNN & HAN using Keras
Stars: ✭ 216 (-1.37%)
Mutual labels: deeplearning
Thor
thor: C++ helper library, for deep learning purpose
Stars: ✭ 197 (-10.05%)
Mutual labels: deeplearning
Release
Deep Reinforcement Learning for de-novo Drug Design
Stars: ✭ 201 (-8.22%)
Mutual labels: deeplearning
Trixi
Manage your machine learning experiments with trixi - modular, reproducible, high fashion. An experiment infrastructure optimized for PyTorch, but flexible enough to work for your framework and your tastes.
Stars: ✭ 211 (-3.65%)
Mutual labels: deeplearning
Halite Ii
Season 2 of @twosigma's artificial intelligence programming challenge
Stars: ✭ 201 (-8.22%)
Mutual labels: deeplearning
Urduhack
An NLP library for the Urdu language. It comes with a lot of battery included features to help you process Urdu data in the easiest way possible.
Stars: ✭ 200 (-8.68%)
Mutual labels: deeplearning
PyTorch models trained on CIFAR-10 dataset
- I modified TorchVision official implementation of popular CNN models, and trained those on CIFAR-10 dataset.
- I changed number of class, filter size, stride, and padding in the the original code so that it works with CIFAR-10.
- I also share the weights of these models, so you can just load the weights and use them.
- The code is highly re-producible and readable by using PyTorch-Lightning.
Statistics of supported models
No. | Model | Val. Acc. | No. Params | Size |
---|---|---|---|---|
1 | vgg11_bn | 92.39% | 28.150 M | 108 MB |
2 | vgg13_bn | 94.22% | 28.334 M | 109 MB |
3 | vgg16_bn | 94.00% | 33.647 M | 129 MB |
4 | vgg19_bn | 93.95% | 38.959 M | 149 MB |
5 | resnet18 | 93.07% | 11.174 M | 43 MB |
6 | resnet34 | 93.34% | 21.282 M | 82 MB |
7 | resnet50 | 93.65% | 23.521 M | 91 MB |
8 | densenet121 | 94.06% | 6.956 M | 28 MB |
9 | densenet161 | 94.07% | 26.483 M | 103 MB |
10 | densenet169 | 94.05% | 12.493 M | 49 MB |
11 | mobilenet_v2 | 93.91% | 2.237 M | 9 MB |
12 | googlenet | 92.85% | 5.491 M | 22 MB |
13 | inception_v3 | 93.74% | 21.640 M | 83 MB |
Details report
Weight and Biases' details report for this project WandB Report
How To Cite
How to use pretrained models
Automatically download and extract the weights from Box (933 MB)
python train.py --download_weights 1
Or use Google Drive backup link (you have to download and extract manually)
Load model and run
from cifar10_models.vgg import vgg11_bn, vgg13_bn, vgg16_bn, vgg19_bn
# Untrained model
my_model = vgg11_bn()
# Pretrained model
my_model = vgg11_bn(pretrained=True)
my_model.eval() # for evaluation
If you use your own images, all models expect data to be in range [0, 1] then normalized by
mean = [0.4914, 0.4822, 0.4465]
std = [0.2471, 0.2435, 0.2616]
How to train models from scratch
Check the train.py
to see all available hyper-parameter choices.
To reproduce the same accuracy use the default hyper-parameters
python train.py --classifier resnet18
How to test pretrained models
python train.py --test_phase 1 --pretrained 1 --classifier resnet18
Output
{'acc/test': tensor(93.0689, device='cuda:0')}
Requirements
Just to use pretrained models
- pytorch = 1.7.0
To train & test
- pytorch = 1.7.0
- torchvision = 0.7.0
- tensorboard = 2.2.1
- pytorch-lightning = 1.1.0
Note that the project description data, including the texts, logos, images, and/or trademarks,
for each open source project belongs to its rightful owner.
If you wish to add or remove any projects, please contact us at [email protected].