PaddleslimPaddleSlim is an open-source library for deep model compression and architecture search.
Stars: ✭ 677 (-66.7%)
BitPackBitPack is a practical tool to efficiently save ultra-low precision/mixed-precision quantized models.
Stars: ✭ 36 (-98.23%)
Kd libA Pytorch Knowledge Distillation library for benchmarking and extending works in the domains of Knowledge Distillation, Pruning, and Quantization.
Stars: ✭ 173 (-91.49%)
Micronetmicronet, a model compression and deploy lib. compression: 1、quantization: quantization-aware-training(QAT), High-Bit(>2b)(DoReFa/Quantization and Training of Neural Networks for Efficient Integer-Arithmetic-Only Inference)、Low-Bit(≤2b)/Ternary and Binary(TWN/BNN/XNOR-Net); post-training-quantization(PTQ), 8-bit(tensorrt); 2、 pruning: normal、regular and group convolutional channel pruning; 3、 group convolution structure; 4、batch-normalization fuse for quantization. deploy: tensorrt, fp32/fp16/int8(ptq-calibration)、op-adapt(upsample)、dynamic_shape
Stars: ✭ 1,232 (-39.4%)
Model OptimizationA toolkit to optimize ML models for deployment for Keras and TensorFlow, including quantization and pruning.
Stars: ✭ 992 (-51.21%)
ZAQ-codeCVPR 2021 : Zero-shot Adversarial Quantization (ZAQ)
Stars: ✭ 59 (-97.1%)
HawqQuantization library for PyTorch. Support low-precision and mixed-precision quantization, with hardware implementation through TVM.
Stars: ✭ 108 (-94.69%)
Tf2An Open Source Deep Learning Inference Engine Based on FPGA
Stars: ✭ 113 (-94.44%)
neural-compressorIntel® Neural Compressor (formerly known as Intel® Low Precision Optimization Tool), targeting to provide unified APIs for network compression technologies, such as low precision quantization, sparsity, pruning, knowledge distillation, across different deep learning frameworks to pursue optimal inference performance.
Stars: ✭ 666 (-67.24%)
Awesome Ml Model CompressionAwesome machine learning model compression research papers, tools, and learning material.
Stars: ✭ 166 (-91.83%)
ATMC[NeurIPS'2019] Shupeng Gui, Haotao Wang, Haichuan Yang, Chen Yu, Zhangyang Wang, Ji Liu, “Model Compression with Adversarial Robustness: A Unified Optimization Framework”
Stars: ✭ 41 (-97.98%)
Awesome Automl And Lightweight ModelsA list of high-quality (newest) AutoML works and lightweight models including 1.) Neural Architecture Search, 2.) Lightweight Structures, 3.) Model Compression, Quantization and Acceleration, 4.) Hyperparameter Optimization, 5.) Automated Feature Engineering.
Stars: ✭ 691 (-66.01%)
LightctrLightweight and Scalable framework that combines mainstream algorithms of Click-Through-Rate prediction based computational DAG, philosophy of Parameter Server and Ring-AllReduce collective communication.
Stars: ✭ 644 (-68.32%)
Keras model compressionModel Compression Based on Geoffery Hinton's Logit Regression Method in Keras applied to MNIST 16x compression over 0.95 percent accuracy.An Implementation of "Distilling the Knowledge in a Neural Network - Geoffery Hinton et. al"
Stars: ✭ 59 (-97.1%)
PaddleclasA treasure chest for image classification powered by PaddlePaddle
Stars: ✭ 625 (-69.26%)
NeuronblocksNLP DNN Toolkit - Building Your NLP DNN Models Like Playing Lego
Stars: ✭ 1,356 (-33.3%)
Jacinto Ai DevkitTraining & Quantization of embedded friendly Deep Learning / Machine Learning / Computer Vision models
Stars: ✭ 49 (-97.59%)
Ghostnet.pytorch[CVPR2020] GhostNet: More Features from Cheap Operations
Stars: ✭ 440 (-78.36%)
DeephashAn Open-Source Package for Deep Learning to Hash (DeepHash)
Stars: ✭ 417 (-79.49%)
Awesome PruningA curated list of neural network pruning resources.
Stars: ✭ 1,017 (-49.98%)
LibimagequantPalette quantization library that powers pngquant and other PNG optimizers
Stars: ✭ 344 (-83.08%)
DistillerNeural Network Distiller by Intel AI Lab: a Python package for neural network compression research. https://intellabs.github.io/distiller
Stars: ✭ 3,760 (+84.95%)
FrostnetFrostNet: Towards Quantization-Aware Network Architecture Search
Stars: ✭ 85 (-95.82%)
Quantization.mxnetSimulate quantization and quantization aware training for MXNet-Gluon models.
Stars: ✭ 42 (-97.93%)
Filter Pruning Geometric MedianFilter Pruning via Geometric Median for Deep Convolutional Neural Networks Acceleration (CVPR 2019 Oral)
Stars: ✭ 338 (-83.37%)
Dsqpytorch implementation of "Differentiable Soft Quantization: Bridging Full-Precision and Low-Bit Neural Networks"
Stars: ✭ 70 (-96.56%)
Pinto model zooA repository that shares tuning results of trained models generated by TensorFlow / Keras. Post-training quantization (Weight Quantization, Integer Quantization, Full Integer Quantization, Float16 Quantization), Quantization-aware training. TensorFlow Lite. OpenVINO. CoreML. TensorFlow.js. TF-TRT. MediaPipe. ONNX. [.tflite,.h5,.pb,saved_model,tfjs,tftrt,mlmodel,.xml/.bin, .onnx]
Stars: ✭ 634 (-68.81%)
NniAn open source AutoML toolkit for automate machine learning lifecycle, including feature engineering, neural architecture search, model compression and hyper-parameter tuning.
Stars: ✭ 10,698 (+426.22%)
Awesome EmdlEmbedded and mobile deep learning research resources
Stars: ✭ 554 (-72.75%)
Ntaggerreference pytorch code for named entity tagging
Stars: ✭ 58 (-97.15%)
AimetAIMET is a library that provides advanced quantization and compression techniques for trained neural network models.
Stars: ✭ 453 (-77.72%)
PngquantLossy PNG compressor — pngquant command based on libimagequant library
Stars: ✭ 4,086 (+100.98%)
BrevitasBrevitas: quantization-aware training in PyTorch
Stars: ✭ 343 (-83.13%)
CompressCompressing Representations for Self-Supervised Learning
Stars: ✭ 43 (-97.88%)
MicroexpnetMicroExpNet: An Extremely Small and Fast Model For Expression Recognition From Frontal Face Images
Stars: ✭ 121 (-94.05%)
Deephash PapersMust-read papers on deep learning to hash (DeepHash)
Stars: ✭ 302 (-85.15%)
Amc[ECCV 2018] AMC: AutoML for Model Compression and Acceleration on Mobile Devices
Stars: ✭ 298 (-85.34%)
PyeprPowerful, automated analysis and design of quantum microwave chips & devices [Energy-Participation Ratio and more]
Stars: ✭ 81 (-96.02%)
Knowledge Distillation PytorchA PyTorch implementation for exploring deep and shallow knowledge distillation (KD) experiments with flexibility
Stars: ✭ 986 (-51.5%)
Soft Filter PruningSoft Filter Pruning for Accelerating Deep Convolutional Neural Networks
Stars: ✭ 291 (-85.69%)
FinnDataflow compiler for QNN inference on FPGAs
Stars: ✭ 284 (-86.03%)
Channel PruningChannel Pruning for Accelerating Very Deep Neural Networks (ICCV'17)
Stars: ✭ 979 (-51.84%)
QkerasQKeras: a quantization deep learning library for Tensorflow Keras
Stars: ✭ 254 (-87.51%)
SaiSDK for TEE AI Stick (includes model training script, inference library, examples)
Stars: ✭ 28 (-98.62%)
sparsifyEasy-to-use UI for automatically sparsifying neural networks and creating sparsification recipes for better inference performance and a smaller footprint
Stars: ✭ 138 (-93.21%)
SViTE[NeurIPS'21] "Chasing Sparsity in Vision Transformers: An End-to-End Exploration" by Tianlong Chen, Yu Cheng, Zhe Gan, Lu Yuan, Lei Zhang, Zhangyang Wang
Stars: ✭ 50 (-97.54%)
MWPToolkitMWPToolkit is an open-source framework for math word problem(MWP) solvers.
Stars: ✭ 67 (-96.7%)