Haq[CVPR 2019, Oral] HAQ: Hardware-Aware Automated Quantization with Mixed Precision
BlueoilBring Deep Learning to small devices
Nlp ArchitectA model library for exploring state-of-the-art deep learning topologies and techniques for optimizing Natural Language Processing neural networks
NncfPyTorch*-based Neural Network Compression Framework for enhanced OpenVINO™ inference
Lq NetsLQ-Nets: Learned Quantization for Highly Accurate and Compact Deep Neural Networks
Pytorch PlaygroundBase pretrained models and datasets in pytorch (MNIST, SVHN, CIFAR10, CIFAR100, STL10, AlexNet, VGG16, VGG19, ResNet, Inception, SqueezeNet)
Kd libA Pytorch Knowledge Distillation library for benchmarking and extending works in the domains of Knowledge Distillation, Pruning, and Quantization.
TerngradTernary Gradients to Reduce Communication in Distributed Deep Learning (TensorFlow)
Zeroq[CVPR'20] ZeroQ: A Novel Zero Shot Quantization Framework
Inq PytorchA PyTorch implementation of "Incremental Network Quantization: Towards Lossless CNNs with Low-Precision Weights"
Awesome Edge Machine LearningA curated list of awesome edge machine learning resources, including research papers, inference engines, challenges, books, meetups and others.
GraffitistGraph Transforms to Quantize and Retrain Deep Neural Nets in TensorFlow
DfqPyTorch implementation of Data Free Quantization Through Weight Equalization and Bias Correction.
Tf2An Open Source Deep Learning Inference Engine Based on FPGA
HawqQuantization library for PyTorch. Support low-precision and mixed-precision quantization, with hardware implementation through TVM.
FrostnetFrostNet: Towards Quantization-Aware Network Architecture Search
PyeprPowerful, automated analysis and design of quantum microwave chips & devices [Energy-Participation Ratio and more]
Micronetmicronet, a model compression and deploy lib. compression: 1、quantization: quantization-aware-training(QAT), High-Bit(>2b)(DoReFa/Quantization and Training of Neural Networks for Efficient Integer-Arithmetic-Only Inference)、Low-Bit(≤2b)/Ternary and Binary(TWN/BNN/XNOR-Net); post-training-quantization(PTQ), 8-bit(tensorrt); 2、 pruning: normal、regular and group convolutional channel pruning; 3、 group convolution structure; 4、batch-normalization fuse for quantization. deploy: tensorrt, fp32/fp16/int8(ptq-calibration)、op-adapt(upsample)、dynamic_shape
VectorsinsearchDice.com repo to accompany the dice.com 'Vectors in Search' talk by Simon Hughes, from the Activate 2018 search conference, and the 'Searching with Vectors' talk from Haystack 2019 (US). Builds upon my conceptual search and semantic search work from 2015
Dsqpytorch implementation of "Differentiable Soft Quantization: Bridging Full-Precision and Low-Bit Neural Networks"
Ntaggerreference pytorch code for named entity tagging
Jacinto Ai DevkitTraining & Quantization of embedded friendly Deep Learning / Machine Learning / Computer Vision models
Quantization.mxnetSimulate quantization and quantization aware training for MXNet-Gluon models.
Model OptimizationA toolkit to optimize ML models for deployment for Keras and TensorFlow, including quantization and pruning.
SaiSDK for TEE AI Stick (includes model training script, inference library, examples)
Awesome Automl And Lightweight ModelsA list of high-quality (newest) AutoML works and lightweight models including 1.) Neural Architecture Search, 2.) Lightweight Structures, 3.) Model Compression, Quantization and Acceleration, 4.) Hyperparameter Optimization, 5.) Automated Feature Engineering.
PaddleslimPaddleSlim is an open-source library for deep model compression and architecture search.
Pinto model zooA repository that shares tuning results of trained models generated by TensorFlow / Keras. Post-training quantization (Weight Quantization, Integer Quantization, Full Integer Quantization, Float16 Quantization), Quantization-aware training. TensorFlow Lite. OpenVINO. CoreML. TensorFlow.js. TF-TRT. MediaPipe. ONNX. [.tflite,.h5,.pb,saved_model,tfjs,tftrt,mlmodel,.xml/.bin, .onnx]
PaddleclasA treasure chest for image classification powered by PaddlePaddle
Awesome EmdlEmbedded and mobile deep learning research resources
AimetAIMET is a library that provides advanced quantization and compression techniques for trained neural network models.
DeephashAn Open-Source Package for Deep Learning to Hash (DeepHash)
PngquantLossy PNG compressor — pngquant command based on libimagequant library
BrevitasBrevitas: quantization-aware training in PyTorch
LibimagequantPalette quantization library that powers pngquant and other PNG optimizers
DistillerNeural Network Distiller by Intel AI Lab: a Python package for neural network compression research. https://intellabs.github.io/distiller
FinnDataflow compiler for QNN inference on FPGAs
QkerasQKeras: a quantization deep learning library for Tensorflow Keras
Pretrained Language ModelPretrained language model and its related optimization techniques developed by Huawei Noah's Ark Lab.
sparsifyEasy-to-use UI for automatically sparsifying neural networks and creating sparsification recipes for better inference performance and a smaller footprint
quantize🎨 Simple color palette quantization using MMCQ
colorquantGo library for color quantization and dithering
aaai17-cdqThe implementation of AAAI-17 paper "Collective Deep Quantization of Efficient Cross-modal Retrieval"
KGySoft.DrawingKGy SOFT Drawing is a library for advanced image, icon and graphics handling.
optimum🏎️ Accelerate training and inference of 🤗 Transformers with easy to use hardware optimization tools
ppqPPL Quantization Tool (PPQ) is a powerful offline neural network quantization tool.