All Projects → MoritzTaylor → maml-rl-tf2

MoritzTaylor / maml-rl-tf2

Licence: MIT License
Implementation of Model-Agnostic Meta-Learning (MAML) applied on Reinforcement Learning problems in TensorFlow 2.

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to maml-rl-tf2

tesp
Implementation of our paper "Meta Reinforcement Learning with Task Embedding and Shared Policy"
Stars: ✭ 28 (+75%)
Mutual labels:  meta-learning, meta-rl, meta-reinforcement
tensorflow-maml
TensorFlow 2.0 implementation of MAML.
Stars: ✭ 79 (+393.75%)
Mutual labels:  maml, meta-learning, tensorflow2
Lottery Ticket Hypothesis-TensorFlow 2
Implementing "The Lottery Ticket Hypothesis" paper by "Jonathan Frankle, Michael Carbin"
Stars: ✭ 28 (+75%)
Mutual labels:  paper, tensorflow2
Epg
Code for the paper "Evolved Policy Gradients"
Stars: ✭ 204 (+1175%)
Mutual labels:  paper, meta-learning
MeTAL
Official PyTorch implementation of "Meta-Learning with Task-Adaptive Loss Function for Few-Shot Learning" (ICCV2021 Oral)
Stars: ✭ 24 (+50%)
Mutual labels:  maml, meta-learning
PAML
Personalizing Dialogue Agents via Meta-Learning
Stars: ✭ 114 (+612.5%)
Mutual labels:  maml, meta-learning
Transferlearning
Transfer learning / domain adaptation / domain generalization / multi-task learning etc. Papers, codes, datasets, applications, tutorials.-迁移学习
Stars: ✭ 8,481 (+52906.25%)
Mutual labels:  paper, meta-learning
FSL-Mate
FSL-Mate: A collection of resources for few-shot learning (FSL).
Stars: ✭ 1,346 (+8312.5%)
Mutual labels:  paper, meta-learning
Tensorflow-YOLACT
Implementation of the paper "YOLACT Real-time Instance Segmentation" in Tensorflow 2
Stars: ✭ 97 (+506.25%)
Mutual labels:  paper, tensorflow2
MetaBIN
[CVPR2021] Meta Batch-Instance Normalization for Generalizable Person Re-Identification
Stars: ✭ 58 (+262.5%)
Mutual labels:  maml, meta-learning
FOCAL-ICLR
Code for FOCAL Paper Published at ICLR 2021
Stars: ✭ 35 (+118.75%)
Mutual labels:  meta-learning, meta-rl
CS330-Stanford-Deep-Multi-Task-and-Meta-Learning
My notes and assignment solutions for Stanford CS330 (Fall 2019 & 2020) Deep Multi-Task and Meta Learning
Stars: ✭ 34 (+112.5%)
Mutual labels:  maml, meta-learning
mliis
Code for meta-learning initializations for image segmentation
Stars: ✭ 21 (+31.25%)
Mutual labels:  maml, meta-learning
maml-tensorflow
This repository implements the paper, Model-Agnostic Meta-Leanring for Fast Adaptation of Deep Networks.
Stars: ✭ 17 (+6.25%)
Mutual labels:  maml, meta-learning
paper-template
Collection of paper latex template for several computer vision related conference.
Stars: ✭ 63 (+293.75%)
Mutual labels:  paper
faceboxes-tensorflow
a tensorflow implement faceboxes
Stars: ✭ 45 (+181.25%)
Mutual labels:  tensorflow2
TF2-RL
Reinforcement learning algorithms implemented for Tensorflow 2.0+ [DQN, DDPG, AE-DDPG, SAC, PPO, Primal-Dual DDPG]
Stars: ✭ 160 (+900%)
Mutual labels:  tensorflow2
nalp
🗣️ NALP is a library that covers Natural Adversarial Language Processing.
Stars: ✭ 17 (+6.25%)
Mutual labels:  tensorflow2
FaceRecognition With FaceNet Android
Face Recognition using the FaceNet model and MLKit on Android.
Stars: ✭ 113 (+606.25%)
Mutual labels:  tensorflow2
DeepBeerInventory-RL
The code for the SRDQN algorithm to train an agent for the beer game problem
Stars: ✭ 27 (+68.75%)
Mutual labels:  paper

Reinforcement Learning with Model-Agnostic Meta-Learning (MAML) in TensorFlow 2 (TF2)

Implementation of Model-Agnostic Meta-Learning (MAML) applied on Reinforcement Learning problems in TensorFlow 2.

This repo is heavily inspired by the original implementation cbfinn/maml_rl (TensorFlow 1.x) as well as the fantastic implementations of Tristan Deleu (tristandeleu/pytorch-maml-rl (PyTorch)) and Jonas Rothfuss (jonasrothfuss/ProMP (TensorFlow 1.x)). I totally recommend to check out all three implementations too.

The original MAML algorithm uses TRPO as optimization method and so far this is also integrated in this version. Tests with 2DNavigation-v0 and HalfCheetahDir-v1 environments yield the same results as the original paper. Better TF2 graph support with tf.function and more variations of MAML (i.e. CAVIA, ProMP, etc.) might be added soon.

Usage

You can use the main.py script in order to train the algorithm with MAML.

python main.py --env-name 2DNavigation-v0 --num-workers 20 --fast-lr 0.1 --max-kl 0.01 --fast-batch-size 20 --meta-batch-size 40 --num-layers 2 --hidden-size 100 --num-batches 500 --gamma 0.99 --tau 1.0 --cg-damping 1e-5 --ls-max-steps 15

To evaluate the trained agent, just run

python experiments.py

Both scripts were tested with Python 3.6.

References

This project is, for the most part, a reproduction of the original implementation cbfinn/maml_rl in TensorFlow 2. The experiments are based on the paper

Chelsea Finn, Pieter Abbeel, and Sergey Levine. Model-agnostic meta-learning for fast adaptation of deep networks. International Conference on Machine Learning (ICML), 2017 [ArXiv]

If you want to cite this paper

@article{DBLP:journals/corr/FinnAL17,
  author    = {Chelsea Finn and Pieter Abbeel and Sergey Levine},
  title     = {Model-{A}gnostic {M}eta-{L}earning for {F}ast {A}daptation of {D}eep {N}etworks},
  journal   = {International Conference on Machine Learning (ICML)},
  year      = {2017},
  url       = {http://arxiv.org/abs/1703.03400}
}
Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].