All Projects → Regularization-Pruning → Similar Projects or Alternatives

141 Open source projects that are alternatives of or similar to Regularization-Pruning

Distiller
Neural Network Distiller by Intel AI Lab: a Python package for neural network compression research. https://intellabs.github.io/distiller
Stars: ✭ 3,760 (+8445.45%)
Mutual labels:  regularization, pruning
torchprune
A research library for pytorch-based neural network pruning, compression, and more.
Stars: ✭ 133 (+202.27%)
Mutual labels:  pruning, filter-pruning
Torch Pruning
A pytorch pruning toolkit for structured neural network pruning and layer dependency maintaining.
Stars: ✭ 193 (+338.64%)
Mutual labels:  pruning, model-compression
DS-Net
(CVPR 2021, Oral) Dynamic Slimmable Network
Stars: ✭ 204 (+363.64%)
Mutual labels:  pruning, model-compression
Soft Filter Pruning
Soft Filter Pruning for Accelerating Deep Convolutional Neural Networks
Stars: ✭ 291 (+561.36%)
Mutual labels:  pruning, model-compression
Paddleslim
PaddleSlim is an open-source library for deep model compression and architecture search.
Stars: ✭ 677 (+1438.64%)
Mutual labels:  pruning, model-compression
Awesome Pruning
A curated list of neural network pruning resources.
Stars: ✭ 1,017 (+2211.36%)
Mutual labels:  pruning, model-compression
torch-model-compression
针对pytorch模型的自动化模型结构分析和修改工具集,包含自动分析模型结构的模型压缩算法库
Stars: ✭ 126 (+186.36%)
Mutual labels:  pruning, model-compression
ATMC
[NeurIPS'2019] Shupeng Gui, Haotao Wang, Haichuan Yang, Chen Yu, Zhangyang Wang, Ji Liu, “Model Compression with Adversarial Robustness: A Unified Optimization Framework”
Stars: ✭ 41 (-6.82%)
Mutual labels:  pruning, model-compression
Micronet
micronet, a model compression and deploy lib. compression: 1、quantization: quantization-aware-training(QAT), High-Bit(>2b)(DoReFa/Quantization and Training of Neural Networks for Efficient Integer-Arithmetic-Only Inference)、Low-Bit(≤2b)/Ternary and Binary(TWN/BNN/XNOR-Net); post-training-quantization(PTQ), 8-bit(tensorrt); 2、 pruning: normal、regular and group convolutional channel pruning; 3、 group convolution structure; 4、batch-normalization fuse for quantization. deploy: tensorrt, fp32/fp16/int8(ptq-calibration)、op-adapt(upsample)、dynamic_shape
Stars: ✭ 1,232 (+2700%)
Mutual labels:  pruning, model-compression
Filter Pruning Geometric Median
Filter Pruning via Geometric Median for Deep Convolutional Neural Networks Acceleration (CVPR 2019 Oral)
Stars: ✭ 338 (+668.18%)
Mutual labels:  pruning, model-compression
Awesome Ml Model Compression
Awesome machine learning model compression research papers, tools, and learning material.
Stars: ✭ 166 (+277.27%)
Mutual labels:  pruning, model-compression
Model Optimization
A toolkit to optimize ML models for deployment for Keras and TensorFlow, including quantization and pruning.
Stars: ✭ 992 (+2154.55%)
Mutual labels:  pruning, model-compression
SViTE
[NeurIPS'21] "Chasing Sparsity in Vision Transformers: An End-to-End Exploration" by Tianlong Chen, Yu Cheng, Zhe Gan, Lu Yuan, Lei Zhang, Zhangyang Wang
Stars: ✭ 50 (+13.64%)
Mutual labels:  pruning, model-compression
Kd lib
A Pytorch Knowledge Distillation library for benchmarking and extending works in the domains of Knowledge Distillation, Pruning, and Quantization.
Stars: ✭ 173 (+293.18%)
Mutual labels:  pruning, model-compression
Awesome Ai Infrastructures
Infrastructures™ for Machine Learning Training/Inference in Production.
Stars: ✭ 223 (+406.82%)
Mutual labels:  pruning, model-compression
GAN-LTH
[ICLR 2021] "GANs Can Play Lottery Too" by Xuxi Chen, Zhenyu Zhang, Yongduo Sui, Tianlong Chen
Stars: ✭ 24 (-45.45%)
Mutual labels:  pruning
Deep-Learning-Specialization-Coursera
Deep Learning Specialization Course by Coursera. Neural Networks, Deep Learning, Hyper Tuning, Regularization, Optimization, Data Processing, Convolutional NN, Sequence Models are including this Course.
Stars: ✭ 75 (+70.45%)
Mutual labels:  regularization
prunnable-layers-pytorch
Prunable nn layers for pytorch.
Stars: ✭ 47 (+6.82%)
Mutual labels:  pruning
pyowl
Ordered Weighted L1 regularization for classification and regression in Python
Stars: ✭ 52 (+18.18%)
Mutual labels:  regularization
Pruning filters for efficient convnets
PyTorch implementation of "Pruning Filters For Efficient ConvNets"
Stars: ✭ 96 (+118.18%)
Mutual labels:  pruning
SReT
Official PyTorch implementation of our ECCV 2022 paper "Sliced Recursive Transformer"
Stars: ✭ 51 (+15.91%)
Mutual labels:  efficient-neural-networks
Statistical-Learning-using-R
This is a Statistical Learning application which will consist of various Machine Learning algorithms and their implementation in R done by me and their in depth interpretation.Documents and reports related to the below mentioned techniques can be found on my Rpubs profile.
Stars: ✭ 27 (-38.64%)
Mutual labels:  regularization
awesome-efficient-gnn
Code and resources on scalable and efficient Graph Neural Networks
Stars: ✭ 498 (+1031.82%)
Mutual labels:  efficient-neural-networks
FisherPruning
Group Fisher Pruning for Practical Network Compression(ICML2021)
Stars: ✭ 127 (+188.64%)
Mutual labels:  pruning
PenaltyFunctions.jl
Julia package of regularization functions for machine learning
Stars: ✭ 25 (-43.18%)
Mutual labels:  regularization
Auto-Compression
Automatic DNN compression tool with various model compression and neural architecture search techniques
Stars: ✭ 19 (-56.82%)
Mutual labels:  model-compression
Structured-Bayesian-Pruning-pytorch
pytorch implementation of Structured Bayesian Pruning
Stars: ✭ 18 (-59.09%)
Mutual labels:  model-compression
deep-compression
Learning both Weights and Connections for Efficient Neural Networks https://arxiv.org/abs/1506.02626
Stars: ✭ 156 (+254.55%)
Mutual labels:  pruning
bert-squeeze
🛠️ Tools for Transformers compression using PyTorch Lightning ⚡
Stars: ✭ 56 (+27.27%)
Mutual labels:  pruning
pyunfold
Iterative unfolding for Python
Stars: ✭ 23 (-47.73%)
Mutual labels:  regularization
manifold mixup
Tensorflow implementation of the Manifold Mixup machine learning research paper
Stars: ✭ 24 (-45.45%)
Mutual labels:  regularization
sparsebn
Software for learning sparse Bayesian networks
Stars: ✭ 41 (-6.82%)
Mutual labels:  regularization
batchnorm-pruning
Rethinking the Smaller-Norm-Less-Informative Assumption in Channel Pruning of Convolution Layers https://arxiv.org/abs/1802.00124
Stars: ✭ 66 (+50%)
Mutual labels:  pruning
hyperstar
Hyperstar: Negative Sampling Improves Hypernymy Extraction Based on Projection Learning.
Stars: ✭ 24 (-45.45%)
Mutual labels:  regularization
introduction-to-machine-learning
A document covering machine learning basics. 🤖📊
Stars: ✭ 17 (-61.36%)
Mutual labels:  regularization
ZAQ-code
CVPR 2021 : Zero-shot Adversarial Quantization (ZAQ)
Stars: ✭ 59 (+34.09%)
Mutual labels:  model-compression
numpy-neuralnet-exercise
Implementation of key concepts of neuralnetwork via numpy
Stars: ✭ 49 (+11.36%)
Mutual labels:  regularization
PyTorch-Deep-Compression
A PyTorch implementation of the iterative pruning method described in Han et. al. (2015)
Stars: ✭ 39 (-11.36%)
Mutual labels:  pruning
fasterai1
FasterAI: A repository for making smaller and faster models with the FastAI library.
Stars: ✭ 34 (-22.73%)
Mutual labels:  pruning
group sparsity
Group Sparsity: The Hinge Between Filter Pruning and Decomposition for Network Compression. CVPR2020.
Stars: ✭ 45 (+2.27%)
Mutual labels:  filter-pruning
SSE-PT
Codes and Datasets for paper RecSys'20 "SSE-PT: Sequential Recommendation Via Personalized Transformer" and NurIPS'19 "Stochastic Shared Embeddings: Data-driven Regularization of Embedding Layers"
Stars: ✭ 103 (+134.09%)
Mutual labels:  regularization
Generalizing-Lottery-Tickets
This repository contains code to replicate the experiments given in NeurIPS 2019 paper "One ticket to win them all: generalizing lottery ticket initializations across datasets and optimizers"
Stars: ✭ 48 (+9.09%)
Mutual labels:  pruning
ESNAC
Learnable Embedding Space for Efficient Neural Architecture Compression
Stars: ✭ 27 (-38.64%)
Mutual labels:  model-compression
neural-compressor
Intel® Neural Compressor (formerly known as Intel® Low Precision Optimization Tool), targeting to provide unified APIs for network compression technologies, such as low precision quantization, sparsity, pruning, knowledge distillation, across different deep learning frameworks to pursue optimal inference performance.
Stars: ✭ 666 (+1413.64%)
Mutual labels:  pruning
SparseRegression.jl
Statistical Models with Regularization in Pure Julia
Stars: ✭ 37 (-15.91%)
Mutual labels:  regularization
pytorch-network-slimming
A package to make do Network Slimming a little easier
Stars: ✭ 40 (-9.09%)
Mutual labels:  pruning
deep-learning-notes
🧠👨‍💻Deep Learning Specialization • Lecture Notes • Lab Assignments
Stars: ✭ 20 (-54.55%)
Mutual labels:  regularization
SAN
[ECCV 2020] Scale Adaptive Network: Learning to Learn Parameterized Classification Networks for Scalable Input Images
Stars: ✭ 41 (-6.82%)
Mutual labels:  efficient-neural-networks
Deeplearning
Python for《Deep Learning》,该书为《深度学习》(花书) 数学推导、原理剖析与源码级别代码实现
Stars: ✭ 4,020 (+9036.36%)
Mutual labels:  regularization
AMP-Regularizer
Code for our paper "Regularizing Neural Networks via Adversarial Model Perturbation", CVPR2021
Stars: ✭ 26 (-40.91%)
Mutual labels:  regularization
FSCNMF
An implementation of "Fusing Structure and Content via Non-negative Matrix Factorization for Embedding Information Networks".
Stars: ✭ 16 (-63.64%)
Mutual labels:  regularization
tulip
Scaleable input gradient regularization
Stars: ✭ 19 (-56.82%)
Mutual labels:  regularization
NaiveNASflux.jl
Your local Flux surgeon
Stars: ✭ 20 (-54.55%)
Mutual labels:  pruning
FastPose
pytorch realtime multi person keypoint estimation
Stars: ✭ 36 (-18.18%)
Mutual labels:  model-compression
Dynamic Model Pruning with Feedback
Implement of Dynamic Model Pruning with Feedback with pytorch
Stars: ✭ 25 (-43.18%)
Mutual labels:  pruning
mixup
speechpro.com/
Stars: ✭ 23 (-47.73%)
Mutual labels:  regularization
Machine-Learning-Andrew-Ng
机器学习-Coursera-吴恩达- python+Matlab代码实现
Stars: ✭ 127 (+188.64%)
Mutual labels:  regularization
L0Learn
Efficient Algorithms for L0 Regularized Learning
Stars: ✭ 74 (+68.18%)
Mutual labels:  regularization
traj-pred-irl
Official implementation codes of "Regularizing neural networks for future trajectory prediction via IRL framework"
Stars: ✭ 23 (-47.73%)
Mutual labels:  regularization
1-60 of 141 similar projects