Keras Yolo2Easy training on custom dataset. Various backends (MobileNet and SqueezeNet) supported. A YOLO demo to detect raccoon run entirely in brower is accessible at https://git.io/vF7vI (not on Windows).
Stars: ✭ 1,693 (+37.42%)
Cn24Convolutional (Patch) Networks for Semantic Segmentation
Stars: ✭ 125 (-89.85%)
LsuvinitReference caffe implementation of LSUV initialization
Stars: ✭ 99 (-91.96%)
Pytorch FcnPyTorch Implementation of Fully Convolutional Networks. (Training code to reproduce the original result is available.)
Stars: ✭ 1,351 (+9.66%)
StringlifierStringlifier is on Opensource ML Library for detecting random strings in raw text. It can be used in sanitising logs, detecting accidentally exposed credentials and as a pre-processing step in unsupervised ML-based analysis of application text data.
Stars: ✭ 85 (-93.1%)
Haq[CVPR 2019, Oral] HAQ: Hardware-Aware Automated Quantization with Mixed Precision
Stars: ✭ 247 (-79.95%)
BlueoilBring Deep Learning to small devices
Stars: ✭ 244 (-80.19%)
Nlp ArchitectA model library for exploring state-of-the-art deep learning topologies and techniques for optimizing Natural Language Processing neural networks
Stars: ✭ 2,768 (+124.68%)
Lq NetsLQ-Nets: Learned Quantization for Highly Accurate and Compact Deep Neural Networks
Stars: ✭ 195 (-84.17%)
Pytorch PlaygroundBase pretrained models and datasets in pytorch (MNIST, SVHN, CIFAR10, CIFAR100, STL10, AlexNet, VGG16, VGG19, ResNet, Inception, SqueezeNet)
Stars: ✭ 2,201 (+78.65%)
TerngradTernary Gradients to Reduce Communication in Distributed Deep Learning (TensorFlow)
Stars: ✭ 168 (-86.36%)
Zeroq[CVPR'20] ZeroQ: A Novel Zero Shot Quantization Framework
Stars: ✭ 150 (-87.82%)
Inq PytorchA PyTorch implementation of "Incremental Network Quantization: Towards Lossless CNNs with Low-Precision Weights"
Stars: ✭ 147 (-88.07%)
Cnn QuantizationQuantization of Convolutional Neural networks.
Stars: ✭ 141 (-88.56%)
Ctranslate2Fast inference engine for OpenNMT models
Stars: ✭ 140 (-88.64%)
GraffitistGraph Transforms to Quantize and Retrain Deep Neural Nets in TensorFlow
Stars: ✭ 135 (-89.04%)
DfqPyTorch implementation of Data Free Quantization Through Weight Equalization and Bias Correction.
Stars: ✭ 125 (-89.85%)
FrostnetFrostNet: Towards Quantization-Aware Network Architecture Search
Stars: ✭ 85 (-93.1%)
PyeprPowerful, automated analysis and design of quantum microwave chips & devices [Energy-Participation Ratio and more]
Stars: ✭ 81 (-93.43%)
PocketflowAn Automatic Model Compression (AutoMC) framework for developing smaller and faster AI applications.
Stars: ✭ 2,672 (+116.88%)
Bert Of Theseus⛵️The official PyTorch implementation for "BERT-of-Theseus: Compressing BERT by Progressive Module Replacing" (EMNLP 2020).
Stars: ✭ 209 (-83.04%)
JfasttextJava interface for fastText
Stars: ✭ 193 (-84.33%)
Mobile IdDeep Face Model Compression
Stars: ✭ 187 (-84.82%)
PruningCode for "Co-Evolutionary Compression for Unpaired Image Translation" (ICCV 2019) and "SCOP: Scientific Control for Reliable Neural Network Pruning" (NeurIPS 2020).
Stars: ✭ 159 (-87.09%)
Amc Models[ECCV 2018] AMC: AutoML for Model Compression and Acceleration on Mobile Devices
Stars: ✭ 154 (-87.5%)
Ld NetEfficient Contextualized Representation: Language Model Pruning for Sequence Labeling
Stars: ✭ 148 (-87.99%)
Yolov3yolov3 by pytorch
Stars: ✭ 142 (-88.47%)
Collaborative DistillationPyTorch code for our CVPR'20 paper "Collaborative Distillation for Ultra-Resolution Universal Style Transfer"
Stars: ✭ 138 (-88.8%)
CondensaProgrammable Neural Network Compression
Stars: ✭ 129 (-89.53%)
MicroexpnetMicroExpNet: An Extremely Small and Fast Model For Expression Recognition From Frontal Face Images
Stars: ✭ 121 (-90.18%)
GhostnetCV backbones including GhostNet, TinyNet and TNT, developed by Huawei Noah's Ark Lab.
Stars: ✭ 1,744 (+41.56%)
NniAn open source AutoML toolkit for automate machine learning lifecycle, including feature engineering, neural architecture search, model compression and hyper-parameter tuning.
Stars: ✭ 10,698 (+768.34%)
NeuronblocksNLP DNN Toolkit - Building Your NLP DNN Models Like Playing Lego
Stars: ✭ 1,356 (+10.06%)
Selecsls PytorchReference ImageNet implementation of SelecSLS CNN architecture proposed in the SIGGRAPH 2020 paper "XNect: Real-time Multi-Person 3D Motion Capture with a Single RGB Camera". The repository also includes code for pruning the model based on implicit sparsity emerging from adaptive gradient descent methods, as detailed in the CVPR 2019 paper "On implicit filter level sparsity in Convolutional Neural Networks".
Stars: ✭ 251 (-79.63%)
SkimcaffeCaffe for Sparse Convolutional Neural Network
Stars: ✭ 230 (-81.33%)
Neuralnetworks.thought ExperimentsObservations and notes to understand the workings of neural network models and other thought experiments using Tensorflow
Stars: ✭ 199 (-83.85%)
HrankPytorch implementation of our CVPR 2020 (Oral) -- HRank: Filter Pruning using High-Rank Feature Map
Stars: ✭ 164 (-86.69%)
Lottery Ticket Hypothesis In PytorchThis repository contains a Pytorch implementation of the paper "The Lottery Ticket Hypothesis: Finding Sparse, Trainable Neural Networks" by Jonathan Frankle and Michael Carbin that can be easily adapted to any model/dataset.
Stars: ✭ 162 (-86.85%)
Cen[NeurIPS 2020] Code release for paper "Deep Multimodal Fusion by Channel Exchanging" (In PyTorch)
Stars: ✭ 112 (-90.91%)
Filter GraftingFilter Grafting for Deep Neural Networks(CVPR 2020)
Stars: ✭ 110 (-91.07%)
MmlsparkSimple and Distributed Machine Learning
Stars: ✭ 2,899 (+135.31%)