All Projects → kalviny → IMTA

kalviny / IMTA

Licence: other
No description or website provided.

Programming Languages

python
139335 projects - #7 most used programming language
shell
77523 projects

Projects that are alternatives of or similar to IMTA

Dl tutorial
Tutorials for deep learning
Stars: ✭ 247 (+550%)
Mutual labels:  deeplearning
gan deeplearning4j
Automatic feature engineering using Generative Adversarial Networks using Deeplearning4j and Apache Spark.
Stars: ✭ 19 (-50%)
Mutual labels:  deeplearning
TensorFlow20-Notes
TensorFlow 2.0 Various notes and tutorials
Stars: ✭ 14 (-63.16%)
Mutual labels:  deeplearning
C3d Pytorch
Pytorch porting of C3D network, with Sports1M weights
Stars: ✭ 251 (+560.53%)
Mutual labels:  deeplearning
Savior
(WIP)The deployment framework aims to provide a simple, lightweight, fast integrated, pipelined deployment framework for algorithm service that ensures reliability, high concurrency and scalability of services.
Stars: ✭ 124 (+226.32%)
Mutual labels:  deeplearning
Baidu-Dog2017
http://js.baidu.com/
Stars: ✭ 37 (-2.63%)
Mutual labels:  deeplearning
Speechbrain.github.io
The SpeechBrain project aims to build a novel speech toolkit fully based on PyTorch. With SpeechBrain users can easily create speech processing systems, ranging from speech recognition (both HMM/DNN and end-to-end), speaker recognition, speech enhancement, speech separation, multi-microphone speech processing, and many others.
Stars: ✭ 242 (+536.84%)
Mutual labels:  deeplearning
googlecodelabs
TPU ile Yapay Sinir Ağlarınızı Çok Daha Hızlı Eğitin
Stars: ✭ 116 (+205.26%)
Mutual labels:  deeplearning
awesome-conformal-prediction
A professionally curated list of awesome Conformal Prediction videos, tutorials, books, papers, PhD and MSc theses, articles and open-source libraries.
Stars: ✭ 998 (+2526.32%)
Mutual labels:  deeplearning
Simplified SqueezeNet
An improved version of SqueezeNet networks https://github.com/DeepScale/SqueezeNet
Stars: ✭ 38 (+0%)
Mutual labels:  deeplearning
Deeplearning Cfn
Distributed Deep Learning on AWS Using CloudFormation (CFN), MXNet and TensorFlow
Stars: ✭ 252 (+563.16%)
Mutual labels:  deeplearning
Dawn Bench Entries
DAWNBench: An End-to-End Deep Learning Benchmark and Competition
Stars: ✭ 254 (+568.42%)
Mutual labels:  deeplearning
Malware-Detection
Deep Learning Based Android Malware Detection Framework
Stars: ✭ 29 (-23.68%)
Mutual labels:  deeplearning
Scan2cad
[CVPR'19] Dataset and code used in the research project Scan2CAD: Learning CAD Model Alignment in RGB-D Scans
Stars: ✭ 249 (+555.26%)
Mutual labels:  deeplearning
smd
Simple mmdetection CPU inference
Stars: ✭ 27 (-28.95%)
Mutual labels:  deeplearning
Awesome Carla
👉 CARLA resources such as tutorial, blog, code and etc https://github.com/carla-simulator/carla
Stars: ✭ 246 (+547.37%)
Mutual labels:  deeplearning
genetic deep learning
No description or website provided.
Stars: ✭ 13 (-65.79%)
Mutual labels:  deeplearning
buildTensorflow
A lightweight deep learning framework made with ❤️
Stars: ✭ 28 (-26.32%)
Mutual labels:  deeplearning
knime-tensorflow
KNIME Deep Learning - Tensorflow Integration
Stars: ✭ 18 (-52.63%)
Mutual labels:  deeplearning
66Days NaturalLanguageProcessing
I am sharing my Journey of 66DaysofData in Natural Language Processing.
Stars: ✭ 127 (+234.21%)
Mutual labels:  deeplearning

IMTA

This repository contains the code (in PyTorch) for "Improved Techniques for Training Adaptive Deep Networks" paper by Hao Li*, Hong Zhang*, Xiaojuan Qi, Ruigang Yang and Gao Huang (* Authors contributed equally).

Introduction

This paper presents three techniques to improve the training efficacy of adaptive deep networks from two aspects: (1) a Gradient Equilibrium algorithm to resolve the conflict of learning of different classifiers; (2) an Inline Subnetwork Collaboration approach and a One-for-all Knowledge Distillation algorithm to enhance the collaboration among classifiers.

Method Overview.

Results

(a) Budgeted prediction results on ImageNet.

(b) Budgeted prediction results on CIFAR-100.

``

Dependencies:

  • Python3
  • PyTorch >= 1.0

Usage

We Provide shell scripts for training an MSDNet on ImageNet with GE, ISC and OFA.

Training an IMTA_MSDNet (block=5, step=4) on ImageNet.

  • Step 1: Training an MSDNet with GE from scratch
    Modify the run_GE.sh to config your path to the ImageNet, your GPU devices and your saving directory. Then run

    bash run_GE.sh
    
  • Step 2: Training the classifiers with ISC and OFA
    Modify the run_IMTA.sh to config your path to the ImageNet, your GPU devices and your saving directory (different from the saving directory of your GE_MSDNet). Please be noted that the MSDNet settings should be exactly the same as your trained GE_MSDNet, and the pretrained directory of IMTA_MSDNet should be the saving directory of your trained GE_MSDNet.

    bash run_IMTA.sh
    
Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].