All Projects → denisyarats → proto

denisyarats / proto

Licence: MIT license
Proto-RL: Reinforcement Learning with Prototypical Representations

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to proto

Drq
DrQ: Data regularized Q
Stars: ✭ 268 (+300%)
Mutual labels:  control, pixel, gym, rl, mujoco
learning-to-drive-in-5-minutes
Implementation of reinforcement learning approach to make a car learn to drive smoothly in minutes
Stars: ✭ 227 (+238.81%)
Mutual labels:  gym, rl, sac, soft-actor-critic
omd
JAX code for the paper "Control-Oriented Model-Based Reinforcement Learning with Implicit Differentiation"
Stars: ✭ 43 (-35.82%)
Mutual labels:  gym, sac, soft-actor-critic
jax-rl
JAX implementations of core Deep RL algorithms
Stars: ✭ 61 (-8.96%)
Mutual labels:  sac, mujoco, soft-actor-critic
mujoco-benchmark
Provide full reinforcement learning benchmark on mujoco environments, including ddpg, sac, td3, pg, a2c, ppo, library
Stars: ✭ 101 (+50.75%)
Mutual labels:  rl, sac, mujoco
Meta-SAC
Auto-tune the Entropy Temperature of Soft Actor-Critic via Metagradient - 7th ICML AutoML workshop 2020
Stars: ✭ 19 (-71.64%)
Mutual labels:  sac, mujoco, soft-actor-critic
Tianshou
An elegant PyTorch deep reinforcement learning library.
Stars: ✭ 4,109 (+6032.84%)
Mutual labels:  rl, sac, mujoco
Deepdrive
Deepdrive is a simulator that allows anyone with a PC to push the state-of-the-art in self-driving
Stars: ✭ 628 (+837.31%)
Mutual labels:  control, gym
Atari
AI research environment for the Atari 2600 games 🤖.
Stars: ✭ 174 (+159.7%)
Mutual labels:  gym, rl
Revisiting-Contrastive-SSL
Revisiting Contrastive Methods for Unsupervised Learning of Visual Representations. [NeurIPS 2021]
Stars: ✭ 81 (+20.9%)
Mutual labels:  representation-learning, unsupervised-learning
VQ-APC
Vector Quantized Autoregressive Predictive Coding (VQ-APC)
Stars: ✭ 34 (-49.25%)
Mutual labels:  representation-learning, unsupervised-learning
FUSION
PyTorch code for NeurIPSW 2020 paper (4th Workshop on Meta-Learning) "Few-Shot Unsupervised Continual Learning through Meta-Examples"
Stars: ✭ 18 (-73.13%)
Mutual labels:  representation-learning, unsupervised-learning
M-NMF
An implementation of "Community Preserving Network Embedding" (AAAI 2017)
Stars: ✭ 119 (+77.61%)
Mutual labels:  representation-learning, unsupervised-learning
Pytorch Reinforce
PyTorch Implementation of REINFORCE for both discrete & continuous control
Stars: ✭ 212 (+216.42%)
Mutual labels:  gym, mujoco
amr
Official adversarial mixup resynthesis repository
Stars: ✭ 31 (-53.73%)
Mutual labels:  representation-learning, unsupervised-learning
Pytorch sac
PyTorch implementation of Soft Actor-Critic (SAC)
Stars: ✭ 174 (+159.7%)
Mutual labels:  gym, mujoco
awesome-graph-self-supervised-learning
Awesome Graph Self-Supervised Learning
Stars: ✭ 805 (+1101.49%)
Mutual labels:  representation-learning, unsupervised-learning
Pytorch-RL-CPP
A Repository with C++ implementations of Reinforcement Learning Algorithms (Pytorch)
Stars: ✭ 73 (+8.96%)
Mutual labels:  gym, mujoco
State-Representation-Learning-An-Overview
Simplified version of "State Representation Learning for Control: An Overview" bibliography
Stars: ✭ 32 (-52.24%)
Mutual labels:  representation-learning, unsupervised-learning
Stable Baselines
Mirror of Stable-Baselines: a fork of OpenAI Baselines, implementations of reinforcement learning algorithms
Stars: ✭ 115 (+71.64%)
Mutual labels:  gym, rl

Proto-RL: Reinforcement Learning with Prototypical Representations

This is a PyTorch implementation of Proto-RL from

Reinforcement Learning with Prototypical Representations by

Denis Yarats, Rob Fergus, Alessandro Lazaric, Lerrel Pinto.

[Paper]

Citation

If you use this repo in your research, please consider citing the paper as follows

@article{yarats2021proto,
    title={Reinforcement Learning with Prototypical Representations},
    author={Denis Yarats and Rob Fergus and Alessandro Lazaric and Lerrel Pinto},
    year={2021},
    eprint={2102.11271},
    archivePrefix={arXiv},
    primaryClass={cs.ML}
}

Requirements

We assume you have access to a gpu that can run CUDA 11. Then, the simplest way to install all required dependencies is to create an anaconda environment by running

conda env create -f conda_env.yml

After the instalation ends you can activate your environment with

conda activate proto

Instructions

In order to pretrain the agent you need to specify the number of task-agnostic environment steps by setting num_expl_steps, after that many steps, the agent will start receving the downstream task reward until it takes num_train_steps in total. For example, to pre-train the Proto-RL agent on Cheetah Run task unsupervisely for 500k environment steps and then train it further with the downstream reward for another 500k steps, you can run:

python train.py env=cheetah_run num_expl_steps=250000 num_train_steps=500000

Note that we divide the number of steps by action repeat, which is set to 2 for all the environments.

This will produce the exp_local folder, where all the outputs are going to be stored including train/eval logs, tensorboard blobs, and evaluation episode videos. To launch tensorboard run

tensorboard --logdir exp_local
Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].