All Projects → for-ai → rl

for-ai / rl

Licence: MIT License
Generic reinforcement learning codebase in TensorFlow

Programming Languages

python
139335 projects - #7 most used programming language
TeX
3793 projects
shell
77523 projects

Projects that are alternatives of or similar to rl

Gymfc
A universal flight control tuning framework
Stars: ✭ 210 (+138.64%)
Mutual labels:  rl
Pytorch-PCGrad
Pytorch reimplementation for "Gradient Surgery for Multi-Task Learning"
Stars: ✭ 179 (+103.41%)
Mutual labels:  rl
proto
Proto-RL: Reinforcement Learning with Prototypical Representations
Stars: ✭ 67 (-23.86%)
Mutual labels:  rl
Learning To Communicate Pytorch
Learning to Communicate with Deep Multi-Agent Reinforcement Learning in PyTorch
Stars: ✭ 236 (+168.18%)
Mutual labels:  rl
gym-rs
OpenAI's Gym written in pure Rust for blazingly fast performance
Stars: ✭ 34 (-61.36%)
Mutual labels:  rl
logrl
Logarithmic Reinforcement Learning
Stars: ✭ 25 (-71.59%)
Mutual labels:  rl
Rl Tutorial Jnrr19
Stable-Baselines tutorial for Journées Nationales de la Recherche en Robotique 2019
Stars: ✭ 204 (+131.82%)
Mutual labels:  rl
Gradient-Samples
Samples for TensorFlow binding for .NET by Lost Tech
Stars: ✭ 53 (-39.77%)
Mutual labels:  rl
revisiting rainbow
Revisiting Rainbow
Stars: ✭ 71 (-19.32%)
Mutual labels:  rl
mdp
Make it easy to specify simple MDPs that are compatible with the OpenAI Gym.
Stars: ✭ 30 (-65.91%)
Mutual labels:  rl
mujoco-benchmark
Provide full reinforcement learning benchmark on mujoco environments, including ddpg, sac, td3, pg, a2c, ppo, library
Stars: ✭ 101 (+14.77%)
Mutual labels:  rl
hra
Hybrid Reward Architecture
Stars: ✭ 76 (-13.64%)
Mutual labels:  rl
Corailed
Unrailed! simulator using C++ with some reinforcement learning and Unrailed! AI using Python with OpenCV
Stars: ✭ 15 (-82.95%)
Mutual labels:  rl
Pytorch Drl
PyTorch implementations of various Deep Reinforcement Learning (DRL) algorithms for both single agent and multi-agent.
Stars: ✭ 233 (+164.77%)
Mutual labels:  rl
RL-code-resources
A collection of Reinforcement Learning GitHub code resources divided by frameworks and environments
Stars: ✭ 51 (-42.05%)
Mutual labels:  rl
Alphazero gomoku
An implementation of the AlphaZero algorithm for Gomoku (also called Gobang or Five in a Row)
Stars: ✭ 2,570 (+2820.45%)
Mutual labels:  rl
neptune-client
📒 Experiment tracking tool and model registry
Stars: ✭ 348 (+295.45%)
Mutual labels:  rl
learning-to-drive-in-5-minutes
Implementation of reinforcement learning approach to make a car learn to drive smoothly in minutes
Stars: ✭ 227 (+157.95%)
Mutual labels:  rl
DeepBeerInventory-RL
The code for the SRDQN algorithm to train an agent for the beer game problem
Stars: ✭ 27 (-69.32%)
Mutual labels:  rl
CartPole
Run OpenAI Gym on a Server
Stars: ✭ 16 (-81.82%)
Mutual labels:  rl

FOR.ai Reinforcement Learning Codebase status DOI Build Status

Modular codebase for reinforcement learning models training, testing and visualization.

Contributors: Bryan M. Li, Alexander Cowen-Rivers, Piotr Kozakowski, David Tao, Siddhartha Rao Kamalakara, Nitarshan Rajkumar, Hariharan Sezhiyan, Sicong Huang, Aidan N. Gomez

Features

Example for recorded envrionment on various RL agents.

MountainCar-v0 Pendulum-v0 VideoPinball-v0 procgen-coinrun-v0
MountainCar-v0 Pendulum-v0 VideoPinball-v0 Tennis-v0

Requirements

It is recommended to install the codebase in a virtual environment (virtualenv or conda).

Quick install

Configure use_gpu and (if on OSX) mac_package_manager (either macports or homebrew) params in setup.sh, then run it as

sh setup.sh

Manual setup

You need to install the following for your system:

Quick Start

# start training
python train.py --sys ... --hparams ... --output_dir ...
# run tensorboard
tensorboard --logdir ...
# test agnet
python train.py --sys ... --hparams ... --output_dir ... --test_only --render

Hyper-parameters

Check available flags with --help, defaults.py for default hyper-parameters, and check hparams/dqn.py agent specific hyper-parameters examples.

  • hparams: Which hparams to use, defined under rl/hparams
  • sys: Which system environment to use.
  • env: Which RL environment to use.
  • output_dir: The directory for model checkpoints and TensorBoard summary.
  • train_steps:, Number of steps to train the agent.
  • test_episodes: Number of episodes to test the agent.
  • eval_episodes: Number of episodes to evaluate the agent.
  • test_only: Test agent without training.
  • copies: Number of independent training/testing runs to do.
  • render: Render game play.
  • record_video: Record game play.
  • num_workers, number of workers.

Documentation

More detailed documentation can be found here.

Contributing

We'd love to accept your contributions to this project. Please feel free to open an issue, or submit a pull request as necessary. Contact us [email protected] for potential collaborations and joining FOR.ai.

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].