All Projects → amzn → Xfer

amzn / Xfer

Licence: apache-2.0
Transfer Learning library for Deep Neural Networks.

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to Xfer

Autogluon
AutoGluon: AutoML for Text, Image, and Tabular Data
Stars: ✭ 3,920 (+2114.69%)
Mutual labels:  transfer-learning, mxnet
Accel Brain Code
The purpose of this repository is to make prototypes as case study in the context of proof of concept(PoC) and research and development(R&D) that I have written in my website. The main research topics are Auto-Encoders in relation to the representation learning, the statistical machine learning for energy-based models, adversarial generation networks(GANs), Deep Reinforcement Learning such as Deep Q-Networks, semi-supervised learning, and neural network language model for natural language processing.
Stars: ✭ 166 (-6.21%)
Mutual labels:  transfer-learning
Learn To Select Data
Code for Learning to select data for transfer learning with Bayesian Optimization
Stars: ✭ 140 (-20.9%)
Mutual labels:  transfer-learning
Transferlearning Tutorial
《迁移学习简明手册》LaTex源码
Stars: ✭ 2,122 (+1098.87%)
Mutual labels:  transfer-learning
Ncnn
ncnn is a high-performance neural network inference framework optimized for the mobile platform
Stars: ✭ 13,376 (+7457.06%)
Mutual labels:  mxnet
Crnn Mxnet Chinese Text Recognition
An implementation of CRNN (CNN+LSTM+warpCTC) on MxNet for chinese text recognition
Stars: ✭ 161 (-9.04%)
Mutual labels:  mxnet
Complete Life Cycle Of A Data Science Project
Complete-Life-Cycle-of-a-Data-Science-Project
Stars: ✭ 140 (-20.9%)
Mutual labels:  transfer-learning
Kashgari
Kashgari is a production-level NLP Transfer learning framework built on top of tf.keras for text-labeling and text-classification, includes Word2Vec, BERT, and GPT2 Language Embedding.
Stars: ✭ 2,235 (+1162.71%)
Mutual labels:  transfer-learning
Coach
Reinforcement Learning Coach by Intel AI Lab enables easy experimentation with state of the art Reinforcement Learning algorithms
Stars: ✭ 2,085 (+1077.97%)
Mutual labels:  mxnet
Transfer Learning Anime
Transfer Learning for Anime Characters Recognition
Stars: ✭ 155 (-12.43%)
Mutual labels:  transfer-learning
Deep Learning Containers
AWS Deep Learning Containers (DLCs) are a set of Docker images for training and serving models in TensorFlow, TensorFlow 2, PyTorch, and MXNet.
Stars: ✭ 152 (-14.12%)
Mutual labels:  mxnet
Djl
An Engine-Agnostic Deep Learning Framework in Java
Stars: ✭ 2,262 (+1177.97%)
Mutual labels:  mxnet
Image keras
Building an image classifier using keras
Stars: ✭ 162 (-8.47%)
Mutual labels:  transfer-learning
Machine Learning Using K8s
Train and Deploy Machine Learning Models on Kubernetes using Amazon EKS
Stars: ✭ 145 (-18.08%)
Mutual labels:  mxnet
Pytorch Retraining
Transfer Learning Shootout for PyTorch's model zoo (torchvision)
Stars: ✭ 167 (-5.65%)
Mutual labels:  transfer-learning
Nas
Neural architecture search(NAS)
Stars: ✭ 140 (-20.9%)
Mutual labels:  mxnet
Pyeco
python implementation of efficient convolution operators for tracking
Stars: ✭ 150 (-15.25%)
Mutual labels:  mxnet
Mobulaop
A Simple & Flexible Cross Framework Operators Toolkit
Stars: ✭ 161 (-9.04%)
Mutual labels:  mxnet
Imgclsmob
Sandbox for training deep learning networks
Stars: ✭ 2,405 (+1258.76%)
Mutual labels:  mxnet
Mx Lsoftmax
mxnet version of Large-Margin Softmax Loss for Convolutional Neural Networks.
Stars: ✭ 175 (-1.13%)
Mutual labels:  mxnet

Xfer

Transfer and meta-learning in Python


Each folder in this repository corresponds to a method or tool for transfer/meta-learning. xfer-ml is a standalone MXNet library (installable with pip) which largely automates deep transfer learning. The rest of the folders contain research code for a novel method in transfer or meta-learning, implemented in a variety of frameworks (not necessarily in MXNet).

In more detail:

  • xfer-ml: A library that allows quick and easy transfer of knowledge stored in deep neural networks implemented in MXNet. xfer-ml can be used with data of arbitrary numeric format, and can be applied to the common cases of image or text data. It can be used as a pipeline that spans from extracting features to training a repurposer. The repurposer is then an object that carries out predictions in the target task. You can also use individual components of the library as part of your own pipeline. For example, you can leverage the feature extractor to extract features from deep neural networks or ModelHandler, which allows for quick building of neural networks, even if you are not an MXNet expert.
  • leap: MXNet implementation of "leap", the meta-gradient path learner (link) by S. Flennerhag, P. G. Moreno, N. Lawrence, A. Damianou, which appeared at ICLR 2019.
  • nn_similarity_index: PyTorch code for comparing trained neural networks using both feature and gradient information. The method is used in the arXiv paper (link) by S. Tang, W. Maddox, C. Dickens, T. Diethe and A. Damianou.
  • finite_ntk: PyTorch implementation of finite width neural tangent kernels from the paper On Transfer Learning with Linearised Neural Networks (link), by W. Maddox, S. Tang, P. G. Moreno, A. G. Wilson, and A. Damianou, which appeared at the NeurIPS MetaLearning Workshop 2019.
  • synthetic_info_bottleneck PyTorch implementation of the Synthetic Information Bottleneck algorithm for few-shot classification on Mini-ImageNet, which is used in paper Empirical Bayes Transductive Meta-Learning with Synthetic Gradients (link) by S. X. Hu, P. G. Moreno, Y. Xiao, X. Shen, G. Obozinski, N. Lawrence and A. Damianou, which appeared at ICLR 2020.
  • var_info_distil PyTorch implementation of the paper Variational Information Distillation for Knowledge Transfer (link) by S. Ahn, S. X. Hu, A. Damianou, N. Lawrence, Z. Dai, which appeared at CVPR 2019.

Navigate to the corresponding folder for more details.

Contributing

You may contribute to the existing projects by reading the individual contribution guidelines in each corresponding folder.

License

The code under this repository is licensed under the Apache 2.0 License.

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].