Pretrained Language ModelPretrained language model and its related optimization techniques developed by Huawei Noah's Ark Lab.
Stars: ✭ 2,033 (+1782.41%)
Awesome Automl And Lightweight ModelsA list of high-quality (newest) AutoML works and lightweight models including 1.) Neural Architecture Search, 2.) Lightweight Structures, 3.) Model Compression, Quantization and Acceleration, 4.) Hyperparameter Optimization, 5.) Automated Feature Engineering.
Stars: ✭ 691 (+539.81%)
Model OptimizationA toolkit to optimize ML models for deployment for Keras and TensorFlow, including quantization and pruning.
Stars: ✭ 992 (+818.52%)
Tf2An Open Source Deep Learning Inference Engine Based on FPGA
Stars: ✭ 113 (+4.63%)
Micronetmicronet, a model compression and deploy lib. compression: 1、quantization: quantization-aware-training(QAT), High-Bit(>2b)(DoReFa/Quantization and Training of Neural Networks for Efficient Integer-Arithmetic-Only Inference)、Low-Bit(≤2b)/Ternary and Binary(TWN/BNN/XNOR-Net); post-training-quantization(PTQ), 8-bit(tensorrt); 2、 pruning: normal、regular and group convolutional channel pruning; 3、 group convolution structure; 4、batch-normalization fuse for quantization. deploy: tensorrt, fp32/fp16/int8(ptq-calibration)、op-adapt(upsample)、dynamic_shape
Stars: ✭ 1,232 (+1040.74%)
Kd libA Pytorch Knowledge Distillation library for benchmarking and extending works in the domains of Knowledge Distillation, Pruning, and Quantization.
Stars: ✭ 173 (+60.19%)
ZAQ-codeCVPR 2021 : Zero-shot Adversarial Quantization (ZAQ)
Stars: ✭ 59 (-45.37%)
PaddleslimPaddleSlim is an open-source library for deep model compression and architecture search.
Stars: ✭ 677 (+526.85%)
BitPackBitPack is a practical tool to efficiently save ultra-low precision/mixed-precision quantized models.
Stars: ✭ 36 (-66.67%)
ATMC[NeurIPS'2019] Shupeng Gui, Haotao Wang, Haichuan Yang, Chen Yu, Zhangyang Wang, Ji Liu, “Model Compression with Adversarial Robustness: A Unified Optimization Framework”
Stars: ✭ 41 (-62.04%)
Awesome EmdlEmbedded and mobile deep learning research resources
Stars: ✭ 554 (+412.96%)
Xxl RpcA high performance, distributed RPC framework.(分布式服务框架XXL-RPC)
Stars: ✭ 493 (+356.48%)
Ghostnet.pytorch[CVPR2020] GhostNet: More Features from Cheap Operations
Stars: ✭ 440 (+307.41%)
PyeprPowerful, automated analysis and design of quantum microwave chips & devices [Energy-Participation Ratio and more]
Stars: ✭ 81 (-25%)
CompressCompressing Representations for Self-Supervised Learning
Stars: ✭ 43 (-60.19%)
DeephashAn Open-Source Package for Deep Learning to Hash (DeepHash)
Stars: ✭ 417 (+286.11%)
BrevitasBrevitas: quantization-aware training in PyTorch
Stars: ✭ 343 (+217.59%)
DistillerNeural Network Distiller by Intel AI Lab: a Python package for neural network compression research. https://intellabs.github.io/distiller
Stars: ✭ 3,760 (+3381.48%)
NniAn open source AutoML toolkit for automate machine learning lifecycle, including feature engineering, neural architecture search, model compression and hyper-parameter tuning.
Stars: ✭ 10,698 (+9805.56%)
VectorsinsearchDice.com repo to accompany the dice.com 'Vectors in Search' talk by Simon Hughes, from the Activate 2018 search conference, and the 'Searching with Vectors' talk from Haystack 2019 (US). Builds upon my conceptual search and semantic search work from 2015
Stars: ✭ 71 (-34.26%)
Channel PruningChannel Pruning for Accelerating Very Deep Neural Networks (ICCV'17)
Stars: ✭ 979 (+806.48%)
Sofa RpcSOFARPC is a high-performance, high-extensibility, production-level Java RPC framework.
Stars: ✭ 3,479 (+3121.3%)
PaddleclasA treasure chest for image classification powered by PaddlePaddle
Stars: ✭ 625 (+478.7%)
Jacinto Ai DevkitTraining & Quantization of embedded friendly Deep Learning / Machine Learning / Computer Vision models
Stars: ✭ 49 (-54.63%)
FrostnetFrostNet: Towards Quantization-Aware Network Architecture Search
Stars: ✭ 85 (-21.3%)
AimetAIMET is a library that provides advanced quantization and compression techniques for trained neural network models.
Stars: ✭ 453 (+319.44%)
Awesome PruningA curated list of neural network pruning resources.
Stars: ✭ 1,017 (+841.67%)
JupiterJupiter是一款性能非常不错的, 轻量级的分布式服务框架
Stars: ✭ 1,372 (+1170.37%)
PngquantLossy PNG compressor — pngquant command based on libimagequant library
Stars: ✭ 4,086 (+3683.33%)
Quantization.mxnetSimulate quantization and quantization aware training for MXNet-Gluon models.
Stars: ✭ 42 (-61.11%)
LibimagequantPalette quantization library that powers pngquant and other PNG optimizers
Stars: ✭ 344 (+218.52%)
Knowledge Distillation PytorchA PyTorch implementation for exploring deep and shallow knowledge distillation (KD) experiments with flexibility
Stars: ✭ 986 (+812.96%)
Filter Pruning Geometric MedianFilter Pruning via Geometric Median for Deep Convolutional Neural Networks Acceleration (CVPR 2019 Oral)
Stars: ✭ 338 (+212.96%)
Sofa HessianAn internal improved version of Hessian powered by Ant Financial.
Stars: ✭ 105 (-2.78%)
SaiSDK for TEE AI Stick (includes model training script, inference library, examples)
Stars: ✭ 28 (-74.07%)
Deephash PapersMust-read papers on deep learning to hash (DeepHash)
Stars: ✭ 302 (+179.63%)
Amc[ECCV 2018] AMC: AutoML for Model Compression and Acceleration on Mobile Devices
Stars: ✭ 298 (+175.93%)
AquvitaeThe Easiest Knowledge Distillation Library for Lightweight Deep Learning
Stars: ✭ 71 (-34.26%)
Soft Filter PruningSoft Filter Pruning for Accelerating Deep Convolutional Neural Networks
Stars: ✭ 291 (+169.44%)
FinnDataflow compiler for QNN inference on FPGAs
Stars: ✭ 284 (+162.96%)
BipointnetThis project is the official implementation of our accepted ICLR 2021 paper BiPointNet: Binary Neural Network for Point Clouds.
Stars: ✭ 27 (-75%)
Dubbogoa golang micro-service framework compatible with alibaba dubbo
Stars: ✭ 258 (+138.89%)
QkerasQKeras: a quantization deep learning library for Tensorflow Keras
Stars: ✭ 254 (+135.19%)
NeuronblocksNLP DNN Toolkit - Building Your NLP DNN Models Like Playing Lego
Stars: ✭ 1,356 (+1155.56%)
Dsqpytorch implementation of "Differentiable Soft Quantization: Bridging Full-Precision and Low-Bit Neural Networks"
Stars: ✭ 70 (-35.19%)
Libimagequant Rustlibimagequant (pngquant) bindings for the Rust language
Stars: ✭ 17 (-84.26%)
sparsifyEasy-to-use UI for automatically sparsifying neural networks and creating sparsification recipes for better inference performance and a smaller footprint
Stars: ✭ 138 (+27.78%)
SViTE[NeurIPS'21] "Chasing Sparsity in Vision Transformers: An End-to-End Exploration" by Tianlong Chen, Yu Cheng, Zhe Gan, Lu Yuan, Lei Zhang, Zhangyang Wang
Stars: ✭ 50 (-53.7%)
quantize🎨 Simple color palette quantization using MMCQ
Stars: ✭ 24 (-77.78%)
Keras model compressionModel Compression Based on Geoffery Hinton's Logit Regression Method in Keras applied to MNIST 16x compression over 0.95 percent accuracy.An Implementation of "Distilling the Knowledge in a Neural Network - Geoffery Hinton et. al"
Stars: ✭ 59 (-45.37%)