All Categories → Machine Learning → quantization

Top 81 quantization open source projects

Haq
[CVPR 2019, Oral] HAQ: Hardware-Aware Automated Quantization with Mixed Precision
Blueoil
Bring Deep Learning to small devices
Nlp Architect
A model library for exploring state-of-the-art deep learning topologies and techniques for optimizing Natural Language Processing neural networks
Nncf
PyTorch*-based Neural Network Compression Framework for enhanced OpenVINO™ inference
Lq Nets
LQ-Nets: Learned Quantization for Highly Accurate and Compact Deep Neural Networks
Pytorch Playground
Base pretrained models and datasets in pytorch (MNIST, SVHN, CIFAR10, CIFAR100, STL10, AlexNet, VGG16, VGG19, ResNet, Inception, SqueezeNet)
Kd lib
A Pytorch Knowledge Distillation library for benchmarking and extending works in the domains of Knowledge Distillation, Pruning, and Quantization.
Terngrad
Ternary Gradients to Reduce Communication in Distributed Deep Learning (TensorFlow)
Awesome Ml Model Compression
Awesome machine learning model compression research papers, tools, and learning material.
Model compression
PyTorch Model Compression
Zeroq
[CVPR'20] ZeroQ: A Novel Zero Shot Quantization Framework
Inq Pytorch
A PyTorch implementation of "Incremental Network Quantization: Towards Lossless CNNs with Low-Precision Weights"
Cnn Quantization
Quantization of Convolutional Neural networks.
Awesome Edge Machine Learning
A curated list of awesome edge machine learning resources, including research papers, inference engines, challenges, books, meetups and others.
Graffitist
Graph Transforms to Quantize and Retrain Deep Neural Nets in TensorFlow
Dfq
PyTorch implementation of Data Free Quantization Through Weight Equalization and Bias Correction.
Tf2
An Open Source Deep Learning Inference Engine Based on FPGA
Hawq
Quantization library for PyTorch. Support low-precision and mixed-precision quantization, with hardware implementation through TVM.
Pyepr
Powerful, automated analysis and design of quantum microwave chips & devices [Energy-Participation Ratio and more]
Micronet
micronet, a model compression and deploy lib. compression: 1、quantization: quantization-aware-training(QAT), High-Bit(>2b)(DoReFa/Quantization and Training of Neural Networks for Efficient Integer-Arithmetic-Only Inference)、Low-Bit(≤2b)/Ternary and Binary(TWN/BNN/XNOR-Net); post-training-quantization(PTQ), 8-bit(tensorrt); 2、 pruning: normal、regular and group convolutional channel pruning; 3、 group convolution structure; 4、batch-normalization fuse for quantization. deploy: tensorrt, fp32/fp16/int8(ptq-calibration)、op-adapt(upsample)、dynamic_shape
Vectorsinsearch
Dice.com repo to accompany the dice.com 'Vectors in Search' talk by Simon Hughes, from the Activate 2018 search conference, and the 'Searching with Vectors' talk from Haystack 2019 (US). Builds upon my conceptual search and semantic search work from 2015
Dsq
pytorch implementation of "Differentiable Soft Quantization: Bridging Full-Precision and Low-Bit Neural Networks"
Ntagger
reference pytorch code for named entity tagging
Jacinto Ai Devkit
Training & Quantization of embedded friendly Deep Learning / Machine Learning / Computer Vision models
Quantization.mxnet
Simulate quantization and quantization aware training for MXNet-Gluon models.
Model Optimization
A toolkit to optimize ML models for deployment for Keras and TensorFlow, including quantization and pruning.
Sai
SDK for TEE AI Stick (includes model training script, inference library, examples)
Libimagequant Rust
libimagequant (pngquant) bindings for the Rust language
Awesome Automl And Lightweight Models
A list of high-quality (newest) AutoML works and lightweight models including 1.) Neural Architecture Search, 2.) Lightweight Structures, 3.) Model Compression, Quantization and Acceleration, 4.) Hyperparameter Optimization, 5.) Automated Feature Engineering.
Paddleslim
PaddleSlim is an open-source library for deep model compression and architecture search.
Pinto model zoo
A repository that shares tuning results of trained models generated by TensorFlow / Keras. Post-training quantization (Weight Quantization, Integer Quantization, Full Integer Quantization, Float16 Quantization), Quantization-aware training. TensorFlow Lite. OpenVINO. CoreML. TensorFlow.js. TF-TRT. MediaPipe. ONNX. [.tflite,.h5,.pb,saved_model,tfjs,tftrt,mlmodel,.xml/.bin, .onnx]
Paddleclas
A treasure chest for image classification powered by PaddlePaddle
Awesome Emdl
Embedded and mobile deep learning research resources
Aimet
AIMET is a library that provides advanced quantization and compression techniques for trained neural network models.
Deephash
An Open-Source Package for Deep Learning to Hash (DeepHash)
Pngquant
Lossy PNG compressor — pngquant command based on libimagequant library
Libimagequant
Palette quantization library that powers pngquant and other PNG optimizers
Distiller
Neural Network Distiller by Intel AI Lab: a Python package for neural network compression research. https://intellabs.github.io/distiller
Deephash Papers
Must-read papers on deep learning to hash (DeepHash)
Finn
Dataflow compiler for QNN inference on FPGAs
Qkeras
QKeras: a quantization deep learning library for Tensorflow Keras
Pretrained Language Model
Pretrained language model and its related optimization techniques developed by Huawei Noah's Ark Lab.
sparsify
Easy-to-use UI for automatically sparsifying neural networks and creating sparsification recipes for better inference performance and a smaller footprint
quantize
🎨 Simple color palette quantization using MMCQ
colorquant
Go library for color quantization and dithering
aaai17-cdq
The implementation of AAAI-17 paper "Collective Deep Quantization of Efficient Cross-modal Retrieval"
U-Net-Fixed-Point-Quantization-for-Medical-Image-Segmentation
Repository containing code for "U-Net Fixed-Point Quantization for Medical Image Segmentation" paper at MICCAI2019
optimum
🏎️ Accelerate training and inference of 🤗 Transformers with easy to use hardware optimization tools
tensorflow-quantization-example
TensorFlow Quantization Example, for TensorFlow Lite
autoencoder based image compression
Autoencoder based image compression: can the learning be quantization independent? https://arxiv.org/abs/1802.09371
ppq
PPL Quantization Tool (PPQ) is a powerful offline neural network quantization tool.
1-60 of 81 quantization projects