All Projects → XinDongol → DNNAC

XinDongol / DNNAC

Licence: other
All about acceleration and compression of Deep Neural Networks

Projects that are alternatives of or similar to DNNAC

Model Optimization
A toolkit to optimize ML models for deployment for Keras and TensorFlow, including quantization and pruning.
Stars: ✭ 992 (+3320.69%)
Mutual labels:  compression, quantization
SSD-Pruning-and-quantization
Pruning and quantization for SSD. Model compression.
Stars: ✭ 19 (-34.48%)
Mutual labels:  compression, quantization
Lq Nets
LQ-Nets: Learned Quantization for Highly Accurate and Compact Deep Neural Networks
Stars: ✭ 195 (+572.41%)
Mutual labels:  compression, quantization
Aimet
AIMET is a library that provides advanced quantization and compression techniques for trained neural network models.
Stars: ✭ 453 (+1462.07%)
Mutual labels:  compression, quantization
Model Quantization
Collections of model quantization algorithms
Stars: ✭ 118 (+306.9%)
Mutual labels:  compression, quantization
Zeroq
[CVPR'20] ZeroQ: A Novel Zero Shot Quantization Framework
Stars: ✭ 150 (+417.24%)
Mutual labels:  compression, quantization
Nncf
PyTorch*-based Neural Network Compression Framework for enhanced OpenVINO™ inference
Stars: ✭ 218 (+651.72%)
Mutual labels:  compression, quantization
supersnappy
Dependency-free and performant Nim Snappy implementation.
Stars: ✭ 55 (+89.66%)
Mutual labels:  compression
FastPose
pytorch realtime multi person keypoint estimation
Stars: ✭ 36 (+24.14%)
Mutual labels:  acceleration
VszLib
7-zip VB6 Helper
Stars: ✭ 35 (+20.69%)
Mutual labels:  compression
salvador
A free, open-source compressor for the ZX0 format
Stars: ✭ 35 (+20.69%)
Mutual labels:  compression
MGARD
MGARD: MultiGrid Adaptive Reduction of Data
Stars: ✭ 16 (-44.83%)
Mutual labels:  compression
SSffmpegVideoOperation
This is a library of FFmpeg for android... 📸 🎞 🚑
Stars: ✭ 261 (+800%)
Mutual labels:  compression
TranscodingStreams.jl
Simple, consistent interfaces for any codec.
Stars: ✭ 71 (+144.83%)
Mutual labels:  compression
memscrimper
Code for the DIMVA 2018 paper: "MemScrimper: Time- and Space-Efficient Storage of Malware Sandbox Memory Dumps"
Stars: ✭ 25 (-13.79%)
Mutual labels:  compression
compressstream-explainer
Compression Streams Explained
Stars: ✭ 22 (-24.14%)
Mutual labels:  compression
pcc geo cnn v2
Improved Deep Point Cloud Geometry Compression
Stars: ✭ 55 (+89.66%)
Mutual labels:  compression
ATMC
[NeurIPS'2019] Shupeng Gui, Haotao Wang, Haichuan Yang, Chen Yu, Zhangyang Wang, Ji Liu, “Model Compression with Adversarial Robustness: A Unified Optimization Framework”
Stars: ✭ 41 (+41.38%)
Mutual labels:  quantization
AGD
[ICML2020] "AutoGAN-Distiller: Searching to Compress Generative Adversarial Networks" by Yonggan Fu, Wuyang Chen, Haotao Wang, Haoran Li, Yingyan Lin, Zhangyang Wang
Stars: ✭ 98 (+237.93%)
Mutual labels:  compression
Stochastic-Quantization
Training Low-bits DNNs with Stochastic Quantization
Stars: ✭ 70 (+141.38%)
Mutual labels:  quantization

DNNAC

All about acceleration and compression of Deep Neural Networks


Quantization

General

  • XNOR-Net: ImageNet Classification Using Binary Convolutional Neural Networks

    A classic paper for binary neural network saying all weights and activation are binarized.

    Implementation: MXNet, Pytorch, Torch (origin)

  • DoReFa-Net: Training Low Bitwidth Convolutional Neural Networks with Low Bitwidth Gradients

    Full stack quantization for weights, activation and gradient.

    Implementation: Tensorpack

  • Deep Learning with Low Precision by Half-wave Gaussian Quantization

    Try to improve expersiveness of quantized activation function.

    Implementation: Caffe (origin)

  • Quantizing deep convolutional networks for efficient inference: A whitepaper

    Non-official technical report of quantization from Google. You can find a lot of technical details about quantization in this paper.

  • Data-Free Quantization through Weight Equalization and Bias Correction

    Implementation: Pytorch

  • Additive Noise Annealing and Approximation Properties of Quantized Neural Networks

    Implementation: Pytorch

  • Towards Learning of Filter-Level Heterogeneous Compression of Convolutional Neural Networks

    find optimal bit-width with NAS

    Implementation: Pytorch

  • Progressive Stochastic Binarization of Deep Networks

    Use power-of-2

    Implementation: TF

  • Trained Quantization Thresholds for Accurate and Efficient Fixed-Point Inference of Deep Neural Networks

    how to find the optimal threshold

    Implementation: TF

  • FAT: Fast Adjustable Threshold for Uniform Neural Network Quantization (Winning Solution on LPIRC-II)

    Implementation: TF

  • Proximal Mean-field for Neural Network Quantization

    Implementation: Pytorch

  • A Survey on Methods and Theories of Quantized Neural Networks

    Nice survey on quantization (up to Dec. 2018)

Binary
  • Balanced Binary Neural Networks with Gated Residual
  • IR-Net: Forward and Backward Information Retention for Highly Accurate Binary Neural Networks

Application-oriented

NLP
  • Differentiable Product Quantization for Embedding Compression

    compress the embedding table with end-to-end learned KD codes via differentiable product quantization (DPQ)

    Implementation: TF

Adversarial
  • Model Compression with Adversarial Robustness: A Unified Optimization Framework

    This paper studies model compression through a different lens: could we compress models without hurting their robustness to adversarial attacks, in addition to maintaining accuracy?

    Implementation: Pytorch

Pruning

  • Learning both Weights and Connections for Efficient Neural Networks

    A very simple way to introduce arbitrary sparisity.

  • Learning Structured Sparsity in Deep Neural Networks

    An united way to introduce structured sparsity.

    Implementation: Caffe

Neural Architecture Search (NAS)

  • Resource
    1. automl.org
  • Partial Channel Connections for Memory-Efficient Differentiable Architecture Search

    Our approach is memory efficient:(i) batch-size is increased to further accelerate the search on CIFAR10, (ii) directly search on ImageNet. Searched on ImageNet, we achieved currently one of, if not only, the best performance on ImageNet (24.2%/7.3%) under the mobile setting! The search process in CIFAR10 only requires 0.1 GPU-days, i.e., ~3 hours on one Nvidia 1080ti.(1.5 hours on one Tesla V100) Implementation: PyTorch (origin)

Others

  • Benchmark Analysis of Representative Deep Neural Network Architectures [IEEE Access, University of Milano-Bicocca]

    This work presents an in-depth analysis of the majority of the deep neural networks (DNNs) proposed in the state of the art for image recognition in terms of GFLOPs, #weights, Top-1 accuacy and so on.

  • Net2Net : Accelerating Learning via Knowledge Transfer

    An interesting way to change the architecture of models while keeping output the same

    Implementation: TF, Pytorch

Embedded System

  • EMDL: Embedded and mobile deep learning research notes

    Embedded and mobile deep learning research notes on Github

Tools

Research

  • slimmable_networks

    An open source framework for slimmable training on tasks of ImageNet classification and COCO detection, which has enabled numerous projects.

  • distiller

    a Python package for neural network compression research

  • QPyTorch

    QPyTorch is a low-precision arithmetic simulation package in PyTorch. It is designed to support researches on low-precision machine learning, especially for researches in low-precision training.

  • Graffitist

    Graffitist is a flexible and scalable framework built on top of TensorFlow to process low-level graph descriptions of deep neural networks (DNNs) for accurate and efficient inference on fixed-point hardware. It comprises of a (growing) library of transforms to apply various neural network compression techniques such as quantization, pruning, and compression. Each transform consists of unique pattern matching and manipulation algorithms that when run sequentially produce an optimized output graph.

Industry

  • dabnn

    dabnn is an accelerated binary neural networks inference framework for mobile platform

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].