All Projects → Pretrained Language Model → Similar Projects or Alternatives

160 Open source projects that are alternatives of or similar to Pretrained Language Model

torch-model-compression
针对pytorch模型的自动化模型结构分析和修改工具集,包含自动分析模型结构的模型压缩算法库
Stars: ✭ 126 (-93.8%)
Mutual labels:  quantization, model-compression
Paddleslim
PaddleSlim is an open-source library for deep model compression and architecture search.
Stars: ✭ 677 (-66.7%)
Mutual labels:  model-compression, quantization
BitPack
BitPack is a practical tool to efficiently save ultra-low precision/mixed-precision quantized models.
Stars: ✭ 36 (-98.23%)
Mutual labels:  quantization, model-compression
Kd lib
A Pytorch Knowledge Distillation library for benchmarking and extending works in the domains of Knowledge Distillation, Pruning, and Quantization.
Stars: ✭ 173 (-91.49%)
Mutual labels:  model-compression, quantization
Micronet
micronet, a model compression and deploy lib. compression: 1、quantization: quantization-aware-training(QAT), High-Bit(>2b)(DoReFa/Quantization and Training of Neural Networks for Efficient Integer-Arithmetic-Only Inference)、Low-Bit(≤2b)/Ternary and Binary(TWN/BNN/XNOR-Net); post-training-quantization(PTQ), 8-bit(tensorrt); 2、 pruning: normal、regular and group convolutional channel pruning; 3、 group convolution structure; 4、batch-normalization fuse for quantization. deploy: tensorrt, fp32/fp16/int8(ptq-calibration)、op-adapt(upsample)、dynamic_shape
Stars: ✭ 1,232 (-39.4%)
Mutual labels:  model-compression, quantization
Model Optimization
A toolkit to optimize ML models for deployment for Keras and TensorFlow, including quantization and pruning.
Stars: ✭ 992 (-51.21%)
Mutual labels:  model-compression, quantization
ZAQ-code
CVPR 2021 : Zero-shot Adversarial Quantization (ZAQ)
Stars: ✭ 59 (-97.1%)
Mutual labels:  quantization, model-compression
Efficient-Computing
Efficient-Computing
Stars: ✭ 474 (-76.68%)
Awesome Ai Infrastructures
Infrastructures™ for Machine Learning Training/Inference in Production.
Stars: ✭ 223 (-89.03%)
Mutual labels:  model-compression, quantization
Hawq
Quantization library for PyTorch. Support low-precision and mixed-precision quantization, with hardware implementation through TVM.
Stars: ✭ 108 (-94.69%)
Mutual labels:  model-compression, quantization
Tf2
An Open Source Deep Learning Inference Engine Based on FPGA
Stars: ✭ 113 (-94.44%)
Mutual labels:  model-compression, quantization
neural-compressor
Intel® Neural Compressor (formerly known as Intel® Low Precision Optimization Tool), targeting to provide unified APIs for network compression technologies, such as low precision quantization, sparsity, pruning, knowledge distillation, across different deep learning frameworks to pursue optimal inference performance.
Stars: ✭ 666 (-67.24%)
Awesome Ml Model Compression
Awesome machine learning model compression research papers, tools, and learning material.
Stars: ✭ 166 (-91.83%)
Mutual labels:  model-compression, quantization
ATMC
[NeurIPS'2019] Shupeng Gui, Haotao Wang, Haichuan Yang, Chen Yu, Zhangyang Wang, Ji Liu, “Model Compression with Adversarial Robustness: A Unified Optimization Framework”
Stars: ✭ 41 (-97.98%)
Mutual labels:  quantization, model-compression
Awesome Automl And Lightweight Models
A list of high-quality (newest) AutoML works and lightweight models including 1.) Neural Architecture Search, 2.) Lightweight Structures, 3.) Model Compression, Quantization and Acceleration, 4.) Hyperparameter Optimization, 5.) Automated Feature Engineering.
Stars: ✭ 691 (-66.01%)
Mutual labels:  model-compression, quantization
Lightctr
Lightweight and Scalable framework that combines mainstream algorithms of Click-Through-Rate prediction based computational DAG, philosophy of Parameter Server and Ring-AllReduce collective communication.
Stars: ✭ 644 (-68.32%)
Mutual labels:  model-compression
Keras model compression
Model Compression Based on Geoffery Hinton's Logit Regression Method in Keras applied to MNIST 16x compression over 0.95 percent accuracy.An Implementation of "Distilling the Knowledge in a Neural Network - Geoffery Hinton et. al"
Stars: ✭ 59 (-97.1%)
Mutual labels:  model-compression
Paddleclas
A treasure chest for image classification powered by PaddlePaddle
Stars: ✭ 625 (-69.26%)
Mutual labels:  quantization
Knowledge Distillation Zoo
Pytorch implementation of various Knowledge Distillation (KD) methods.
Stars: ✭ 514 (-74.72%)
Mutual labels:  model-compression
Neuronblocks
NLP DNN Toolkit - Building Your NLP DNN Models Like Playing Lego
Stars: ✭ 1,356 (-33.3%)
Mutual labels:  model-compression
Jacinto Ai Devkit
Training & Quantization of embedded friendly Deep Learning / Machine Learning / Computer Vision models
Stars: ✭ 49 (-97.59%)
Mutual labels:  quantization
Ghostnet.pytorch
[CVPR2020] GhostNet: More Features from Cheap Operations
Stars: ✭ 440 (-78.36%)
Mutual labels:  model-compression
Deephash
An Open-Source Package for Deep Learning to Hash (DeepHash)
Stars: ✭ 417 (-79.49%)
Mutual labels:  quantization
Awesome Pruning
A curated list of neural network pruning resources.
Stars: ✭ 1,017 (-49.98%)
Mutual labels:  model-compression
Data Efficient Model Compression
Data Efficient Model Compression
Stars: ✭ 380 (-81.31%)
Mutual labels:  model-compression
Libimagequant
Palette quantization library that powers pngquant and other PNG optimizers
Stars: ✭ 344 (-83.08%)
Mutual labels:  quantization
Distiller
Neural Network Distiller by Intel AI Lab: a Python package for neural network compression research. https://intellabs.github.io/distiller
Stars: ✭ 3,760 (+84.95%)
Mutual labels:  quantization
Frostnet
FrostNet: Towards Quantization-Aware Network Architecture Search
Stars: ✭ 85 (-95.82%)
Mutual labels:  quantization
Quantization.mxnet
Simulate quantization and quantization aware training for MXNet-Gluon models.
Stars: ✭ 42 (-97.93%)
Mutual labels:  quantization
Filter Pruning Geometric Median
Filter Pruning via Geometric Median for Deep Convolutional Neural Networks Acceleration (CVPR 2019 Oral)
Stars: ✭ 338 (-83.37%)
Mutual labels:  model-compression
Dsq
pytorch implementation of "Differentiable Soft Quantization: Bridging Full-Precision and Low-Bit Neural Networks"
Stars: ✭ 70 (-96.56%)
Mutual labels:  quantization
Pinto model zoo
A repository that shares tuning results of trained models generated by TensorFlow / Keras. Post-training quantization (Weight Quantization, Integer Quantization, Full Integer Quantization, Float16 Quantization), Quantization-aware training. TensorFlow Lite. OpenVINO. CoreML. TensorFlow.js. TF-TRT. MediaPipe. ONNX. [.tflite,.h5,.pb,saved_model,tfjs,tftrt,mlmodel,.xml/.bin, .onnx]
Stars: ✭ 634 (-68.81%)
Mutual labels:  quantization
Nni
An open source AutoML toolkit for automate machine learning lifecycle, including feature engineering, neural architecture search, model compression and hyper-parameter tuning.
Stars: ✭ 10,698 (+426.22%)
Mutual labels:  model-compression
Awesome Emdl
Embedded and mobile deep learning research resources
Stars: ✭ 554 (-72.75%)
Mutual labels:  quantization
Ntagger
reference pytorch code for named entity tagging
Stars: ✭ 58 (-97.15%)
Mutual labels:  quantization
Aimet
AIMET is a library that provides advanced quantization and compression techniques for trained neural network models.
Stars: ✭ 453 (-77.72%)
Mutual labels:  quantization
Model Quantization
Collections of model quantization algorithms
Stars: ✭ 118 (-94.2%)
Mutual labels:  quantization
Knowledge Distillation Papers
knowledge distillation papers
Stars: ✭ 422 (-79.24%)
Mutual labels:  model-compression
Awesome Knowledge Distillation
Awesome Knowledge-Distillation. 分类整理的知识蒸馏paper(2014-2021)。
Stars: ✭ 1,031 (-49.29%)
Mutual labels:  model-compression
Pngquant
Lossy PNG compressor — pngquant command based on libimagequant library
Stars: ✭ 4,086 (+100.98%)
Mutual labels:  quantization
Trained Ternary Quantization
Reducing the size of convolutional neural networks
Stars: ✭ 90 (-95.57%)
Mutual labels:  quantization
Brevitas
Brevitas: quantization-aware training in PyTorch
Stars: ✭ 343 (-83.13%)
Mutual labels:  quantization
Compress
Compressing Representations for Self-Supervised Learning
Stars: ✭ 43 (-97.88%)
Mutual labels:  model-compression
Microexpnet
MicroExpNet: An Extremely Small and Fast Model For Expression Recognition From Frontal Face Images
Stars: ✭ 121 (-94.05%)
Mutual labels:  model-compression
Deephash Papers
Must-read papers on deep learning to hash (DeepHash)
Stars: ✭ 302 (-85.15%)
Mutual labels:  quantization
Amc
[ECCV 2018] AMC: AutoML for Model Compression and Acceleration on Mobile Devices
Stars: ✭ 298 (-85.34%)
Mutual labels:  model-compression
Model Compression Papers
Papers for deep neural network compression and acceleration
Stars: ✭ 296 (-85.44%)
Mutual labels:  model-compression
Pyepr
Powerful, automated analysis and design of quantum microwave chips & devices [Energy-Participation Ratio and more]
Stars: ✭ 81 (-96.02%)
Mutual labels:  quantization
Knowledge Distillation Pytorch
A PyTorch implementation for exploring deep and shallow knowledge distillation (KD) experiments with flexibility
Stars: ✭ 986 (-51.5%)
Mutual labels:  model-compression
Soft Filter Pruning
Soft Filter Pruning for Accelerating Deep Convolutional Neural Networks
Stars: ✭ 291 (-85.69%)
Mutual labels:  model-compression
Finn
Dataflow compiler for QNN inference on FPGAs
Stars: ✭ 284 (-86.03%)
Mutual labels:  quantization
Channel Pruning
Channel Pruning for Accelerating Very Deep Neural Networks (ICCV'17)
Stars: ✭ 979 (-51.84%)
Mutual labels:  model-compression
Qkeras
QKeras: a quantization deep learning library for Tensorflow Keras
Stars: ✭ 254 (-87.51%)
Mutual labels:  quantization
Knowledge distillation via TF2.0
The codes for recent knowledge distillation algorithms and benchmark results via TF2.0 low-level API
Stars: ✭ 87 (-95.72%)
Mutual labels:  knowledge-distillation
Sai
SDK for TEE AI Stick (includes model training script, inference library, examples)
Stars: ✭ 28 (-98.62%)
Mutual labels:  quantization
sparsify
Easy-to-use UI for automatically sparsifying neural networks and creating sparsification recipes for better inference performance and a smaller footprint
Stars: ✭ 138 (-93.21%)
Mutual labels:  quantization
SViTE
[NeurIPS'21] "Chasing Sparsity in Vision Transformers: An End-to-End Exploration" by Tianlong Chen, Yu Cheng, Zhe Gan, Lu Yuan, Lei Zhang, Zhangyang Wang
Stars: ✭ 50 (-97.54%)
Mutual labels:  model-compression
Training extensions
Trainable models and NN optimization tools
Stars: ✭ 857 (-57.85%)
Mutual labels:  quantization
model optimizer
Model optimizer used in Adlik.
Stars: ✭ 22 (-98.92%)
Mutual labels:  knowledge-distillation
MWPToolkit
MWPToolkit is an open-source framework for math word problem(MWP) solvers.
Stars: ✭ 67 (-96.7%)
Mutual labels:  pretrained-language-model
1-60 of 160 similar projects