All Projects → ChangyWen → wolpertinger_ddpg

ChangyWen / wolpertinger_ddpg

Licence: other
Wolpertinger Training with DDPG (Pytorch), Deep Reinforcement Learning in Large Discrete Action Spaces. Multi-GPU/Singer-GPU/CPU compatible.

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to wolpertinger ddpg

Pytorch Rl
This repository contains model-free deep reinforcement learning algorithms implemented in Pytorch
Stars: ✭ 394 (+795.45%)
Mutual labels:  deep-reinforcement-learning, gym, ddpg
Deep-Reinforcement-Learning-With-Python
Master classic RL, deep RL, distributional RL, inverse RL, and more using OpenAI Gym and TensorFlow with extensive Math
Stars: ✭ 222 (+404.55%)
Mutual labels:  deep-reinforcement-learning, ddpg
Deep-rl-mxnet
Mxnet implementation of Deep Reinforcement Learning papers, such as DQN, PG, DDPG, PPO
Stars: ✭ 26 (-40.91%)
Mutual labels:  deep-reinforcement-learning, ddpg
pytorch-gym
Implementation of the Deep Deterministic Policy Gradient(DDPG) in bullet Gym using pytorch
Stars: ✭ 39 (-11.36%)
Mutual labels:  gym, ddpg
omd
JAX code for the paper "Control-Oriented Model-Based Reinforcement Learning with Implicit Differentiation"
Stars: ✭ 43 (-2.27%)
Mutual labels:  deep-reinforcement-learning, gym
pytorch-distributed
Ape-X DQN & DDPG with pytorch & tensorboard
Stars: ✭ 98 (+122.73%)
Mutual labels:  deep-reinforcement-learning, ddpg
deep rl acrobot
TensorFlow A2C to solve Acrobot, with synchronized parallel environments
Stars: ✭ 32 (-27.27%)
Mutual labels:  deep-reinforcement-learning, ddpg
Pytorch sac
PyTorch implementation of Soft Actor-Critic (SAC)
Stars: ✭ 174 (+295.45%)
Mutual labels:  deep-reinforcement-learning, gym
deeprl-continuous-control
Learning Continuous Control in Deep Reinforcement Learning
Stars: ✭ 14 (-68.18%)
Mutual labels:  deep-reinforcement-learning, ddpg
motion-planner-reinforcement-learning
End to end motion planner using Deep Deterministic Policy Gradient (DDPG) in gazebo
Stars: ✭ 99 (+125%)
Mutual labels:  deep-reinforcement-learning, ddpg
Deep-Reinforcement-Learning-for-Automated-Stock-Trading-Ensemble-Strategy-ICAIF-2020
Live Trading. Please star.
Stars: ✭ 1,251 (+2743.18%)
Mutual labels:  deep-reinforcement-learning, ddpg
Pytorch Drl
PyTorch implementations of various Deep Reinforcement Learning (DRL) algorithms for both single agent and multi-agent.
Stars: ✭ 233 (+429.55%)
Mutual labels:  deep-reinforcement-learning, ddpg
Deeprl
Modularized Implementation of Deep RL Algorithms in PyTorch
Stars: ✭ 2,640 (+5900%)
Mutual labels:  deep-reinforcement-learning, ddpg
deep-rl-quadcopter
Implementation of Deep Deterministic Policy Gradients (DDPG) to teach a Quadcopter How to Fly!
Stars: ✭ 17 (-61.36%)
Mutual labels:  deep-reinforcement-learning, ddpg
Naf Tensorflow
"Continuous Deep Q-Learning with Model-based Acceleration" in TensorFlow
Stars: ✭ 192 (+336.36%)
Mutual labels:  deep-reinforcement-learning, gym
Rainy
☔ Deep RL agents with PyTorch☔
Stars: ✭ 39 (-11.36%)
Mutual labels:  deep-reinforcement-learning, ddpg
DDPG
End to End Mobile Robot Navigation using DDPG (Continuous Control with Deep Reinforcement Learning) based on Tensorflow + Gazebo
Stars: ✭ 41 (-6.82%)
Mutual labels:  deep-reinforcement-learning, ddpg
Deep Reinforcement Learning Algorithms
31 projects in the framework of Deep Reinforcement Learning algorithms: Q-learning, DQN, PPO, DDPG, TD3, SAC, A2C and others. Each project is provided with a detailed training log.
Stars: ✭ 167 (+279.55%)
Mutual labels:  deep-reinforcement-learning, ddpg
Machine Learning Is All You Need
🔥🌟《Machine Learning 格物志》: ML + DL + RL basic codes and notes by sklearn, PyTorch, TensorFlow, Keras & the most important, from scratch!💪 This repository is ALL You Need!
Stars: ✭ 173 (+293.18%)
Mutual labels:  deep-reinforcement-learning, ddpg
reinforcement learning ppo rnd
Deep Reinforcement Learning by using Proximal Policy Optimization and Random Network Distillation in Tensorflow 2 and Pytorch with some explanation
Stars: ✭ 33 (-25%)
Mutual labels:  deep-reinforcement-learning, gym

Wolpertinger Training with DDPG (Pytorch, Multi-GPU/single-GPU/CPU)

Overview

Pytorch version of Wolpertinger Training with DDPG (paper: Deep Reinforcement Learning in Large Discrete Action Spaces).
The code is compatible with training in multi-GPU, single-GPU or CPU.
It is also compatible with both continuous and discrete control of OpenAI gym.
In continuous case, I discretize the action space to use wolpertinger-DDPG training algorithm.

Dependencies

  • python 3.6.8
  • torch 1.1.0
  • OpenAI gym
    • If you get an RunTimeError:NotImplementedError in ActionWrapper.step while training with gym, replace your gym/core.py file with core.py in openai/gym.
  • pyflann
    • This is the library (FLANN, Muja & Lowe, 2014) with approximate nearest-neighbor methods allowed for logarithmic-time lookup complexity relative to the number of actions. However, the python binding of FLANN (pyflann) is written for python 2 and is no longer maintained.
    • To use this package, please put the whole directory pyflann into your (virtual) python environment.
    • Please refer to pyflann for a more detailed instruction if needed.

Usage

  • In Pendulum-v0 (continuous control), discretize the continuous action space to a discrete action spaces with 200000 actions.
    python main.py --env 'Pendulum-v0' --max-actions 200000
  • In CartPole-v1 (discrete control), --max-actions is not needed.
    python main.py --env 'CartPole-v1'
  • To use CPU only:
    python main.py --gpu-ids -1
  • To use single-GPU only:
    python main.py --gpu-ids 0 --gpu-nums 1
  • To use multi-GPU (e.g., use GPU-0 and GPU-1):
    python main.py --gpu-ids 0 1 --gpu-nums 2

Result

  • Please refer to output for the trained models and training log.
    • Pendulum-v0: a gym environment with continuous action space.
    • CartPole-v1: a gym environment with discrete action space.

Project Reference

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].