All Projects → hereismari → tensorflow-maml

hereismari / tensorflow-maml

Licence: Apache-2.0 license
TensorFlow 2.0 implementation of MAML.

Programming Languages

Jupyter Notebook
11667 projects

Projects that are alternatives of or similar to tensorflow-maml

maml-rl-tf2
Implementation of Model-Agnostic Meta-Learning (MAML) applied on Reinforcement Learning problems in TensorFlow 2.
Stars: ✭ 16 (-79.75%)
Mutual labels:  maml, meta-learning, tensorflow2
MeTAL
Official PyTorch implementation of "Meta-Learning with Task-Adaptive Loss Function for Few-Shot Learning" (ICCV2021 Oral)
Stars: ✭ 24 (-69.62%)
Mutual labels:  maml, meta-learning
mliis
Code for meta-learning initializations for image segmentation
Stars: ✭ 21 (-73.42%)
Mutual labels:  maml, meta-learning
CS330-Stanford-Deep-Multi-Task-and-Meta-Learning
My notes and assignment solutions for Stanford CS330 (Fall 2019 & 2020) Deep Multi-Task and Meta Learning
Stars: ✭ 34 (-56.96%)
Mutual labels:  maml, meta-learning
PAML
Personalizing Dialogue Agents via Meta-Learning
Stars: ✭ 114 (+44.3%)
Mutual labels:  maml, meta-learning
MetaBIN
[CVPR2021] Meta Batch-Instance Normalization for Generalizable Person Re-Identification
Stars: ✭ 58 (-26.58%)
Mutual labels:  maml, meta-learning
maml-tensorflow
This repository implements the paper, Model-Agnostic Meta-Leanring for Fast Adaptation of Deep Networks.
Stars: ✭ 17 (-78.48%)
Mutual labels:  maml, meta-learning
TF2DeepFloorplan
TF2 Deep FloorPlan Recognition using a Multi-task Network with Room-boundary-Guided Attention. Enable tensorboard, quantization, flask, tflite, docker, github actions and google colab.
Stars: ✭ 98 (+24.05%)
Mutual labels:  tensorflow2
DataAugmentationTF
Implementation of modern data augmentation techniques in TensorFlow 2.x to be used in your training pipeline.
Stars: ✭ 35 (-55.7%)
Mutual labels:  tensorflow2
geotrellis-server
Tools for building raster processing and display services
Stars: ✭ 65 (-17.72%)
Mutual labels:  maml
Awesome Real World Rl
Great resources for making Reinforcement Learning work in Real Life situations. Papers,projects and more.
Stars: ✭ 234 (+196.2%)
Mutual labels:  meta-learning
LibFewShot
LibFewShot: A Comprehensive Library for Few-shot Learning.
Stars: ✭ 629 (+696.2%)
Mutual labels:  meta-learning
AiSpace
AiSpace: Better practices for deep learning model development and deployment For Tensorflow 2.0
Stars: ✭ 28 (-64.56%)
Mutual labels:  tensorflow2
tensorflow 2.0 tutorial
tensorflow 2.0 正式版实用教程/tutorial
Stars: ✭ 48 (-39.24%)
Mutual labels:  tensorflow2
TFLite-ModelMaker-EfficientDet-Colab-Hands-On
TensorFlow Lite Model Makerで物体検出を行うハンズオン用資料です(Hands-on for object detection with TensorFlow Lite Model Maker)
Stars: ✭ 15 (-81.01%)
Mutual labels:  tensorflow2
pyradox
State of the Art Neural Networks for Deep Learning
Stars: ✭ 61 (-22.78%)
Mutual labels:  tensorflow2
Mil
Code for "One-Shot Visual Imitation Learning via Meta-Learning"
Stars: ✭ 254 (+221.52%)
Mutual labels:  meta-learning
tesp
Implementation of our paper "Meta Reinforcement Learning with Task Embedding and Shared Policy"
Stars: ✭ 28 (-64.56%)
Mutual labels:  meta-learning
checkmate
Training neural networks in TensorFlow 2.0 with 5x less memory
Stars: ✭ 116 (+46.84%)
Mutual labels:  tensorflow2
transformer
Build English-Vietnamese machine translation with ProtonX Transformer. :D
Stars: ✭ 41 (-48.1%)
Mutual labels:  tensorflow2

Reproduction of MAML using TensorFlow 2.0.

This reproduction is highly influenced by the pytorch reproduction by Adrien Lucas Effot available at Paper repro: Deep Metalearning using “MAML” and “Reptile”.

MAML | Neural Net

alt-text-1 alt-text-2

MAML paper

https://arxiv.org/abs/1703.03400

Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks Chelsea Finn, Pieter Abbeel, Sergey Levine

We propose an algorithm for meta-learning that is model-agnostic, in the sense that it is compatible with any model trained with gradient descent and applicable to a variety of different learning problems, including classification, regression, and reinforcement learning. The goal of meta-learning is to train a model on a variety of learning tasks, such that it can solve new learning tasks using only a small number of training samples. In our approach, the parameters of the model are explicitly trained such that a small number of gradient steps with a small amount of training data from a new task will produce good generalization performance on that task. In effect, our method trains the model to be easy to fine-tune. We demonstrate that this approach leads to state-of-the-art performance on two few-shot image classification benchmarks, produces good results on few-shot regression, and accelerates fine-tuning for policy gradient reinforcement learning with neural network policies.


image.png

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].