All Projects → ChanChiChoi → Awesome Model Compression

ChanChiChoi / Awesome Model Compression

papers about model compression

Projects that are alternatives of or similar to Awesome Model Compression

Knowledge Distillation Zoo
Pytorch implementation of various Knowledge Distillation (KD) methods.
Stars: ✭ 514 (+331.93%)
Mutual labels:  model-compression
Compress
Compressing Representations for Self-Supervised Learning
Stars: ✭ 43 (-63.87%)
Mutual labels:  model-compression
Neuronblocks
NLP DNN Toolkit - Building Your NLP DNN Models Like Playing Lego
Stars: ✭ 1,356 (+1039.5%)
Mutual labels:  model-compression
Paddleslim
PaddleSlim is an open-source library for deep model compression and architecture search.
Stars: ✭ 677 (+468.91%)
Mutual labels:  model-compression
Knowledge Distillation Pytorch
A PyTorch implementation for exploring deep and shallow knowledge distillation (KD) experiments with flexibility
Stars: ✭ 986 (+728.57%)
Mutual labels:  model-compression
Awesome Knowledge Distillation
Awesome Knowledge-Distillation. 分类整理的知识蒸馏paper(2014-2021)。
Stars: ✭ 1,031 (+766.39%)
Mutual labels:  model-compression
Knowledge Distillation Papers
knowledge distillation papers
Stars: ✭ 422 (+254.62%)
Mutual labels:  model-compression
Hawq
Quantization library for PyTorch. Support low-precision and mixed-precision quantization, with hardware implementation through TVM.
Stars: ✭ 108 (-9.24%)
Mutual labels:  model-compression
Model Optimization
A toolkit to optimize ML models for deployment for Keras and TensorFlow, including quantization and pruning.
Stars: ✭ 992 (+733.61%)
Mutual labels:  model-compression
Micronet
micronet, a model compression and deploy lib. compression: 1、quantization: quantization-aware-training(QAT), High-Bit(>2b)(DoReFa/Quantization and Training of Neural Networks for Efficient Integer-Arithmetic-Only Inference)、Low-Bit(≤2b)/Ternary and Binary(TWN/BNN/XNOR-Net); post-training-quantization(PTQ), 8-bit(tensorrt); 2、 pruning: normal、regular and group convolutional channel pruning; 3、 group convolution structure; 4、batch-normalization fuse for quantization. deploy: tensorrt, fp32/fp16/int8(ptq-calibration)、op-adapt(upsample)、dynamic_shape
Stars: ✭ 1,232 (+935.29%)
Mutual labels:  model-compression
Awesome Automl And Lightweight Models
A list of high-quality (newest) AutoML works and lightweight models including 1.) Neural Architecture Search, 2.) Lightweight Structures, 3.) Model Compression, Quantization and Acceleration, 4.) Hyperparameter Optimization, 5.) Automated Feature Engineering.
Stars: ✭ 691 (+480.67%)
Mutual labels:  model-compression
Channel Pruning
Channel Pruning for Accelerating Very Deep Neural Networks (ICCV'17)
Stars: ✭ 979 (+722.69%)
Mutual labels:  model-compression
Keras model compression
Model Compression Based on Geoffery Hinton's Logit Regression Method in Keras applied to MNIST 16x compression over 0.95 percent accuracy.An Implementation of "Distilling the Knowledge in a Neural Network - Geoffery Hinton et. al"
Stars: ✭ 59 (-50.42%)
Mutual labels:  model-compression
Lightctr
Lightweight and Scalable framework that combines mainstream algorithms of Click-Through-Rate prediction based computational DAG, philosophy of Parameter Server and Ring-AllReduce collective communication.
Stars: ✭ 644 (+441.18%)
Mutual labels:  model-compression
Nni
An open source AutoML toolkit for automate machine learning lifecycle, including feature engineering, neural architecture search, model compression and hyper-parameter tuning.
Stars: ✭ 10,698 (+8889.92%)
Mutual labels:  model-compression
Ghostnet.pytorch
[CVPR2020] GhostNet: More Features from Cheap Operations
Stars: ✭ 440 (+269.75%)
Mutual labels:  model-compression
Awesome Pruning
A curated list of neural network pruning resources.
Stars: ✭ 1,017 (+754.62%)
Mutual labels:  model-compression
Tf2
An Open Source Deep Learning Inference Engine Based on FPGA
Stars: ✭ 113 (-5.04%)
Mutual labels:  model-compression
Ghostnet
CV backbones including GhostNet, TinyNet and TNT, developed by Huawei Noah's Ark Lab.
Stars: ✭ 1,744 (+1365.55%)
Mutual labels:  model-compression
Aquvitae
The Easiest Knowledge Distillation Library for Lightweight Deep Learning
Stars: ✭ 71 (-40.34%)
Mutual labels:  model-compression

awesome-model-compression

this collecting the papers (main from arxiv.org) about Model compression:
Structure;
Distillation;
Binarization;
Quantization;
Pruning;
Low Rank.

also, some papers and links collected from below, they are all awesome resources:


1990

1993

  • Hassibi, Babak, and David G. Stork. Second order derivatives for network pruning: Optimal brain surgeon .[C]Advances in neural information processing systems. 1993.
  • J. L. Holi and J. N. Hwang. [Finite precision error analysis of neural network hardware implementations]. In Ijcnn-91- Seattle International Joint Conference on Neural Networks, pages 519–525 vol.1, 1993.

1995

1997

1998

2000

2001

2006

2011

2012

  • D. Hammerstrom. [A vlsi architecture for highperformance, low-cost, on-chip learning]. In IJCNN International Joint Conference on Neural Networks, pages 537– 544 vol.2, 2012.

2013

2014

2015

2016

2017

2018

2019


Projects

  • NVIDIA TensorRT:  Programmable Inference Accelerator;  
  • Tencent/PocketFlow:  An Automatic Model Compression (AutoMC) framework for developing smaller and faster AI applications;
  • dmlc/tvm:  Open deep learning compiler stack for cpu, gpu and specialized accelerators;
  • Tencent/ncnn:  ncnn is a high-performance neural network inference framework optimized for the mobile platform;
  • pytorch/glow:  Compiler for Neural Network hardware accelerators;
  • NervanaSystems/neon:  Intel® Nervana™ reference deep learning framework committed to best performance on all hardware;
  • NervanaSystems/distiller:  Neural Network Distiller by Intel AI Lab: a Python package for neural network compression research;
  • MUSCO - framework for model compression using tensor decompositions (PyTorch)
  • OAID/Tengine:  Tengine is a lite, high performance, modular inference engine for embedded device;
  • fpeder/espresso:  Efficient forward propagation for BCNNs;
  • Tensorflow lite:  TensorFlow Lite is an open source deep learning framework for on-device inference.;  
  • Core ML:  Reduce the storage used by the Core ML model inside your app bundle;
  • pytorch-tensor-decompositions:  PyTorch implementation of [1412.6553] and [1511.06530] tensor decomposition methods for convolutional layers;
  • tensorflow/quantize:  
  • mxnet/quantization:  This folder contains examples of quantizing a FP32 model with Intel® MKL-DNN or CUDNN.
  • TensoRT4-Example:  
  • NAF-tensorflow:  "Continuous Deep Q-Learning with Model-based Acceleration" in TensorFlow;
  • Mayo - deep learning framework with fine- and coarse-grained pruning, network slimming, and quantization methods
  • Keras compressor - compression using low-rank approximations, SVD for matrices, Tucker for tensors.
  • Caffe compressor K-means based quantization

others

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].