All Projects → ChenDRAG → mujoco-benchmark

ChenDRAG / mujoco-benchmark

Licence: other
Provide full reinforcement learning benchmark on mujoco environments, including ddpg, sac, td3, pg, a2c, ppo, library

Projects that are alternatives of or similar to mujoco-benchmark

Tianshou
An elegant PyTorch deep reinforcement learning library.
Stars: ✭ 4,109 (+3968.32%)
Mutual labels:  rl, ddpg, sac, drl, mujoco, ppo
Rainy
☔ Deep RL agents with PyTorch☔
Stars: ✭ 39 (-61.39%)
Mutual labels:  ddpg, sac, ppo
Pytorch Drl
PyTorch implementations of various Deep Reinforcement Learning (DRL) algorithms for both single agent and multi-agent.
Stars: ✭ 233 (+130.69%)
Mutual labels:  rl, ddpg, ppo
Minimalrl
Implementations of basic RL algorithms with minimal lines of codes! (pytorch based)
Stars: ✭ 2,051 (+1930.69%)
Mutual labels:  ddpg, sac, ppo
Torchrl
Pytorch Implementation of Reinforcement Learning Algorithms ( Soft Actor Critic(SAC)/ DDPG / TD3 /DQN / A2C/ PPO / TRPO)
Stars: ✭ 90 (-10.89%)
Mutual labels:  ddpg, mujoco, ppo
Deep-Reinforcement-Learning-With-Python
Master classic RL, deep RL, distributional RL, inverse RL, and more using OpenAI Gym and TensorFlow with extensive Math
Stars: ✭ 222 (+119.8%)
Mutual labels:  ddpg, sac, ppo
model-free-algorithms
TD3, SAC, IQN, Rainbow, PPO, Ape-X and etc. in TF1.x
Stars: ✭ 56 (-44.55%)
Mutual labels:  ddpg, sac, ppo
LWDRLC
Lightweight deep RL Libraray for continuous control.
Stars: ✭ 14 (-86.14%)
Mutual labels:  ddpg, sac, ppo
TF2-RL
Reinforcement learning algorithms implemented for Tensorflow 2.0+ [DQN, DDPG, AE-DDPG, SAC, PPO, Primal-Dual DDPG]
Stars: ✭ 160 (+58.42%)
Mutual labels:  ddpg, sac, ppo
ReinforcementLearningZoo.jl
juliareinforcementlearning.org/
Stars: ✭ 46 (-54.46%)
Mutual labels:  ddpg, sac, ppo
ElegantRL
Scalable and Elastic Deep Reinforcement Learning Using PyTorch. Please star. 🔥
Stars: ✭ 2,074 (+1953.47%)
Mutual labels:  ddpg, sac, ppo
Mushroom Rl
Python library for Reinforcement Learning.
Stars: ✭ 442 (+337.62%)
Mutual labels:  rl, ddpg, mujoco
Lagom
lagom: A PyTorch infrastructure for rapid prototyping of reinforcement learning algorithms.
Stars: ✭ 364 (+260.4%)
Mutual labels:  ddpg, mujoco, ppo
proto
Proto-RL: Reinforcement Learning with Prototypical Representations
Stars: ✭ 67 (-33.66%)
Mutual labels:  rl, sac, mujoco
Reinforcement Learning With Tensorflow
Simple Reinforcement learning tutorials, 莫烦Python 中文AI教学
Stars: ✭ 6,948 (+6779.21%)
Mutual labels:  ddpg, ppo
Run Skeleton Run
Reason8.ai PyTorch solution for NIPS RL 2017 challenge
Stars: ✭ 83 (-17.82%)
Mutual labels:  ddpg, ppo
Elegantrl
Lightweight, efficient and stable implementations of deep reinforcement learning algorithms using PyTorch.
Stars: ✭ 575 (+469.31%)
Mutual labels:  ddpg, ppo
Machin
Reinforcement learning library(framework) designed for PyTorch, implements DQN, DDPG, A2C, PPO, SAC, MADDPG, A3C, APEX, IMPALA ...
Stars: ✭ 145 (+43.56%)
Mutual labels:  ddpg, ppo
Ros2learn
ROS 2 enabled Machine Learning algorithms
Stars: ✭ 119 (+17.82%)
Mutual labels:  rl, ppo
Deep Reinforcement Learning For Automated Stock Trading Ensemble Strategy Icaif 2020
Deep Reinforcement Learning for Automated Stock Trading: An Ensemble Strategy. ICAIF 2020. Please star.
Stars: ✭ 518 (+412.87%)
Mutual labels:  ddpg, ppo

This repo only servers as a link to Tianshou's benchmark of Mujoco environments. Latest benchmark is maintained under thu-ml/tianshou. See full benchmark here.

Keywords: deep reinforcement learning, pytorch, mujoco, benchmark, performances, Tianshou, baseline

Tianshou's Mujoco Benchmark

We benchmarked Tianshou algorithm implementations in 9 out of 13 environments from the MuJoCo Gym task suite.

For each supported algorithm and supported mujoco environments, we provide:

  • Default hyperparameters used for benchmark and scripts to reproduce the benchmark;
  • A comparison of performance (or code level details) with other open source implementations or classic papers;
  • Graphs and raw data that can be used for research purposes;
  • Log details obtained during training;
  • Pretrained agents;
  • Some hints on how to tune the algorithm.

Supported algorithms are listed below:

Example benchmark

SAC

Environment Tianshou SpinningUp (Pytorch) SAC paper
Ant 5850.2±475.7 ~3980 ~3720
HalfCheetah 12138.8±1049.3 ~11520 ~10400
Hopper 3542.2±51.5 ~3150 ~3370
Walker2d 5007.0±251.5 ~4250 ~3740
Swimmer 44.4±0.5 ~41.7 N
Humanoid 5488.5±81.2 N ~5200
Reacher -2.6±0.2 N N
InvertedPendulum 1000.0±0.0 N N
InvertedDoublePendulum 9359.5±0.4 N N

Other resources

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].