Model OptimizationA toolkit to optimize ML models for deployment for Keras and TensorFlow, including quantization and pruning.
Stars: ✭ 992 (+561.33%)
AimetAIMET is a library that provides advanced quantization and compression techniques for trained neural network models.
Stars: ✭ 453 (+202%)
PaddleslimPaddleSlim is an open-source library for deep model compression and architecture search.
Stars: ✭ 677 (+351.33%)
Ntaggerreference pytorch code for named entity tagging
Stars: ✭ 58 (-61.33%)
Awesome Ml Model CompressionAwesome machine learning model compression research papers, tools, and learning material.
Stars: ✭ 166 (+10.67%)
Micronetmicronet, a model compression and deploy lib. compression: 1、quantization: quantization-aware-training(QAT), High-Bit(>2b)(DoReFa/Quantization and Training of Neural Networks for Efficient Integer-Arithmetic-Only Inference)、Low-Bit(≤2b)/Ternary and Binary(TWN/BNN/XNOR-Net); post-training-quantization(PTQ), 8-bit(tensorrt); 2、 pruning: normal、regular and group convolutional channel pruning; 3、 group convolution structure; 4、batch-normalization fuse for quantization. deploy: tensorrt, fp32/fp16/int8(ptq-calibration)、op-adapt(upsample)、dynamic_shape
Stars: ✭ 1,232 (+721.33%)
bert-squeeze🛠️ Tools for Transformers compression using PyTorch Lightning ⚡
Stars: ✭ 56 (-62.67%)
sparsifyEasy-to-use UI for automatically sparsifying neural networks and creating sparsification recipes for better inference performance and a smaller footprint
Stars: ✭ 138 (-8%)
NncfPyTorch*-based Neural Network Compression Framework for enhanced OpenVINO™ inference
Stars: ✭ 218 (+45.33%)
neural-compressorIntel® Neural Compressor (formerly known as Intel® Low Precision Optimization Tool), targeting to provide unified APIs for network compression technologies, such as low precision quantization, sparsity, pruning, knowledge distillation, across different deep learning frameworks to pursue optimal inference performance.
Stars: ✭ 666 (+344%)
sparsezooNeural network model repository for highly sparse and sparse-quantized models with matching sparsification recipes
Stars: ✭ 264 (+76%)
Awesome Edge Machine LearningA curated list of awesome edge machine learning resources, including research papers, inference engines, challenges, books, meetups and others.
Stars: ✭ 139 (-7.33%)
Awesome EmdlEmbedded and mobile deep learning research resources
Stars: ✭ 554 (+269.33%)
DistillerNeural Network Distiller by Intel AI Lab: a Python package for neural network compression research. https://intellabs.github.io/distiller
Stars: ✭ 3,760 (+2406.67%)
Kd libA Pytorch Knowledge Distillation library for benchmarking and extending works in the domains of Knowledge Distillation, Pruning, and Quantization.
Stars: ✭ 173 (+15.33%)
ATMC[NeurIPS'2019] Shupeng Gui, Haotao Wang, Haichuan Yang, Chen Yu, Zhangyang Wang, Ji Liu, “Model Compression with Adversarial Robustness: A Unified Optimization Framework”
Stars: ✭ 41 (-72.67%)
Libimagequant Rustlibimagequant (pngquant) bindings for the Rust language
Stars: ✭ 17 (-88.67%)
Pytorch PruningPyTorch Implementation of [1611.06440] Pruning Convolutional Neural Networks for Resource Efficient Inference
Stars: ✭ 740 (+393.33%)
Awesome Automl And Lightweight ModelsA list of high-quality (newest) AutoML works and lightweight models including 1.) Neural Architecture Search, 2.) Lightweight Structures, 3.) Model Compression, Quantization and Acceleration, 4.) Hyperparameter Optimization, 5.) Automated Feature Engineering.
Stars: ✭ 691 (+360.67%)
Pinto model zooA repository that shares tuning results of trained models generated by TensorFlow / Keras. Post-training quantization (Weight Quantization, Integer Quantization, Full Integer Quantization, Float16 Quantization), Quantization-aware training. TensorFlow Lite. OpenVINO. CoreML. TensorFlow.js. TF-TRT. MediaPipe. ONNX. [.tflite,.h5,.pb,saved_model,tfjs,tftrt,mlmodel,.xml/.bin, .onnx]
Stars: ✭ 634 (+322.67%)
VectorsinsearchDice.com repo to accompany the dice.com 'Vectors in Search' talk by Simon Hughes, from the Activate 2018 search conference, and the 'Searching with Vectors' talk from Haystack 2019 (US). Builds upon my conceptual search and semantic search work from 2015
Stars: ✭ 71 (-52.67%)
PaddleclasA treasure chest for image classification powered by PaddlePaddle
Stars: ✭ 625 (+316.67%)
Tf Keras SurgeonPruning and other network surgery for trained TF.Keras models.
Stars: ✭ 25 (-83.33%)
PyeprPowerful, automated analysis and design of quantum microwave chips & devices [Energy-Participation Ratio and more]
Stars: ✭ 81 (-46%)
DfqPyTorch implementation of Data Free Quantization Through Weight Equalization and Bias Correction.
Stars: ✭ 125 (-16.67%)
DeephashAn Open-Source Package for Deep Learning to Hash (DeepHash)
Stars: ✭ 417 (+178%)
BrevitasBrevitas: quantization-aware training in PyTorch
Stars: ✭ 343 (+128.67%)
Adventures In Tensorflow LiteThis repository contains notebooks that show the usage of TensorFlow Lite for quantizing deep neural networks.
Stars: ✭ 79 (-47.33%)
Ctranslate2Fast inference engine for OpenNMT models
Stars: ✭ 140 (-6.67%)
LibimagequantPalette quantization library that powers pngquant and other PNG optimizers
Stars: ✭ 344 (+129.33%)
Dsqpytorch implementation of "Differentiable Soft Quantization: Bridging Full-Precision and Low-Bit Neural Networks"
Stars: ✭ 70 (-53.33%)
PngquantLossy PNG compressor — pngquant command based on libimagequant library
Stars: ✭ 4,086 (+2624%)
Cen[NeurIPS 2020] Code release for paper "Deep Multimodal Fusion by Channel Exchanging" (In PyTorch)
Stars: ✭ 112 (-25.33%)
GraspCode for "Picking Winning Tickets Before Training by Preserving Gradient Flow" https://openreview.net/pdf?id=SkgsACVKPH
Stars: ✭ 58 (-61.33%)
Inq PytorchA PyTorch implementation of "Incremental Network Quantization: Towards Lossless CNNs with Low-Precision Weights"
Stars: ✭ 147 (-2%)
Tf2An Open Source Deep Learning Inference Engine Based on FPGA
Stars: ✭ 113 (-24.67%)
Keras SurgeonPruning and other network surgery for trained Keras models.
Stars: ✭ 339 (+126%)
Filter Pruning Geometric MedianFilter Pruning via Geometric Median for Deep Convolutional Neural Networks Acceleration (CVPR 2019 Oral)
Stars: ✭ 338 (+125.33%)
Jacinto Ai DevkitTraining & Quantization of embedded friendly Deep Learning / Machine Learning / Computer Vision models
Stars: ✭ 49 (-67.33%)
Deephash PapersMust-read papers on deep learning to hash (DeepHash)
Stars: ✭ 302 (+101.33%)
Filter GraftingFilter Grafting for Deep Neural Networks(CVPR 2020)
Stars: ✭ 110 (-26.67%)
DelvePyTorch and Keras model training and layer saturation monitor
Stars: ✭ 49 (-67.33%)
Soft Filter PruningSoft Filter Pruning for Accelerating Deep Convolutional Neural Networks
Stars: ✭ 291 (+94%)
FinnDataflow compiler for QNN inference on FPGAs
Stars: ✭ 284 (+89.33%)
Awesome PruningA curated list of neural network pruning resources.
Stars: ✭ 1,017 (+578%)
QkerasQKeras: a quantization deep learning library for Tensorflow Keras
Stars: ✭ 254 (+69.33%)
SIGIR2021 ConureOne Person, One Model, One World: Learning Continual User Representation without Forgetting
Stars: ✭ 23 (-84.67%)
GraffitistGraph Transforms to Quantize and Retrain Deep Neural Nets in TensorFlow
Stars: ✭ 135 (-10%)
HawqQuantization library for PyTorch. Support low-precision and mixed-precision quantization, with hardware implementation through TVM.
Stars: ✭ 108 (-28%)
Quantization.mxnetSimulate quantization and quantization aware training for MXNet-Gluon models.
Stars: ✭ 42 (-72%)
TextPrunerA PyTorch-based model pruning toolkit for pre-trained language models
Stars: ✭ 94 (-37.33%)
SViTE[NeurIPS'21] "Chasing Sparsity in Vision Transformers: An End-to-End Exploration" by Tianlong Chen, Yu Cheng, Zhe Gan, Lu Yuan, Lei Zhang, Zhangyang Wang
Stars: ✭ 50 (-66.67%)
quantize🎨 Simple color palette quantization using MMCQ
Stars: ✭ 24 (-84%)