All Projects → Eli-YiLi → ScaleNet

Eli-YiLi / ScaleNet

Licence: other
Data-Driven Neuron Allocation for Scale Aggregation Networks

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to ScaleNet

HRFormer
This is an official implementation of our NeurIPS 2021 paper "HRFormer: High-Resolution Transformer for Dense Prediction".
Stars: ✭ 357 (+573.58%)
Mutual labels:  classification
focalloss
Focal Loss of multi-classification in tensorflow
Stars: ✭ 75 (+41.51%)
Mutual labels:  classification
tensorflow-image-classificator
HTTP service for Tensorflow images classification model
Stars: ✭ 15 (-71.7%)
Mutual labels:  classification
InstantDL
InstantDL: An easy and convenient deep learning pipeline for image segmentation and classification
Stars: ✭ 33 (-37.74%)
Mutual labels:  classification
napkinXC
Extremely simple and fast extreme multi-class and multi-label classifiers.
Stars: ✭ 38 (-28.3%)
Mutual labels:  classification
sklearn-audio-classification
An in-depth analysis of audio classification on the RAVDESS dataset. Feature engineering, hyperparameter optimization, model evaluation, and cross-validation with a variety of ML techniques and MLP
Stars: ✭ 31 (-41.51%)
Mutual labels:  classification
kaggle-human-protein-atlas-image-classification
Kaggle 2018 @ Human Protein Atlas Image Classification
Stars: ✭ 34 (-35.85%)
Mutual labels:  classification
Focal Loss Keras
Multi-class classification with focal loss for imbalanced datasets
Stars: ✭ 76 (+43.4%)
Mutual labels:  classification
eeg-rsenet
Motor Imagery EEG Signal Classification Using Random Subspace Ensemble Network
Stars: ✭ 24 (-54.72%)
Mutual labels:  classification
snomed-owl-toolkit
The official SNOMED CT OWL Toolkit. OWL conversion, classification and authoring support.
Stars: ✭ 58 (+9.43%)
Mutual labels:  classification
Dimensionality-reduction-and-classification-on-Hyperspectral-Images-Using-Python
In this repository, You can find the files which implement dimensionality reduction on the hyperspectral image(Indian Pines) with classification.
Stars: ✭ 63 (+18.87%)
Mutual labels:  classification
XED
XED multilingual emotion datasets
Stars: ✭ 34 (-35.85%)
Mutual labels:  classification
ProbQA
Probabilistic question-asking system: the program asks, the users answer. The minimal goal of the program is to identify what the user needs (a target), even if the user is not aware of the existence of such a thing/product/service.
Stars: ✭ 43 (-18.87%)
Mutual labels:  classification
nlp classification
Implementing nlp papers relevant to classification with PyTorch, gluonnlp
Stars: ✭ 224 (+322.64%)
Mutual labels:  classification
sp segmenter
Superpixel-based semantic segmentation, with object pose estimation and tracking. Provided as a ROS package.
Stars: ✭ 33 (-37.74%)
Mutual labels:  classification
pyts-repro
A repository to compare the performance between the algorithms implemented in pyts and the performance reported in the literature
Stars: ✭ 15 (-71.7%)
Mutual labels:  classification
Opensource OBIA processing chain
An open-source semi-automated processing chain for urban OBIA classification.
Stars: ✭ 75 (+41.51%)
Mutual labels:  classification
textlytics
Text processing library for sentiment analysis and related tasks
Stars: ✭ 25 (-52.83%)
Mutual labels:  classification
hmm
A Hidden Markov Model implemented in Javascript
Stars: ✭ 29 (-45.28%)
Mutual labels:  classification
neuralBlack
A Multi-Class Brain Tumor Classifier using Convolutional Neural Network with 99% Accuracy achieved by applying the method of Transfer Learning using Python and Pytorch Deep Learning Framework
Stars: ✭ 36 (-32.08%)
Mutual labels:  classification

ScaleNet

By Yi Li, Zhanghui Kuang, Yimin Chen, Wayne Zhang

SenseTime.

Table of Contents

  1. Introduction
  2. Citation
  3. Approach
  4. Trained models
  5. Experiments
  6. GPU time

Introduction

This is a PyTorch implementation of Data-Driven Neuron Allocation for Scale Aggregation Networks.(CVPR2019) with pretrained models.

Citation

If you use these models in your research, please cite:

@inproceedings{Li2019ScaleNet,
    title={Data-Driven Neuron Allocation for Scale Aggregation Networks},
    author={Li, Yi and Kuang, Zhanghui and Chen, Yimin and Zhang, Wayne},
    booktitle={CVPR},
    year={2019}
}

Approach

Figure 1: architecture of ScaleNet-50.

Figure 2: scale aggregation block.

Trained models

Model Top-1 err. Top-5 err.
ScaleNet-50-light 22.80 6.57
ScaleNet-50 22.02 6.05
ScaleNet-101 20.82 5.42
ScaleNet-152 20.06 5.18

Pytorch:

from pytorch.scalenet import *
model = scalenet50(structure_path='structures/scalenet50.json', ckpt=None) # train from stratch
model = scalenet50(structure_path='structures/scalenet50.json', ckpt='weights/scalenet50.pth') # load pretrained model

The weights are available on BaiduYun with extract code: f1c5

Unlike the paper, we used better training settings: increase the epochs to 120 and replace multi-step learning rate by cosine learning rate.

Experiments

Figure 3: experiments on imagenet classification.

Figure 4: experiments on ms-coco detection.

GPU time

Model Top-1 err. FLOPs(10^9) GPU time(ms)
ResNet-50 24.02 4.1 95
SE-ResNet-50 23.29 4.1 98
ResNeXt-50 22.2 4.2 147
ScaleNet-50 22.2 3.8 93

TensorFlow: (empty models of ResNet, SE-ResNet, ResNeXt, ScaleNet for speed test)

python3 tensorflow/test_speed.py scale|res|se|next

All networks were tested using Tensorflow with GTX 1060 GPU and i7 CPU at batch size 16 and image side 224 on 1000 runs.

Some static-graph frameworks like Tensorflow and TensorRT execute multi-branch models in parallel, while Pytorch and Caffe do not. So we suggest to deploy ScaleNets on Tensorflow and TensorRT.

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].