All Projects → flyyufelix → C51 Ddqn Keras

flyyufelix / C51 Ddqn Keras

Licence: mit
C51-DDQN in Keras

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to C51 Ddqn Keras

Reinforcement Learning
🤖 Implements of Reinforcement Learning algorithms.
Stars: ✭ 104 (-9.57%)
Mutual labels:  reinforcement-learning
Pairstrade Fyp 2019
We tested 3 approaches for Pair Trading: distance, cointegration and reinforcement learning approach.
Stars: ✭ 109 (-5.22%)
Mutual labels:  reinforcement-learning
Doom Net Pytorch
Reinforcement learning models in ViZDoom environment
Stars: ✭ 113 (-1.74%)
Mutual labels:  reinforcement-learning
Aws Robomaker Sample Application Deepracer
Use AWS RoboMaker and demonstrate running a simulation which trains a reinforcement learning (RL) model to drive a car around a track
Stars: ✭ 105 (-8.7%)
Mutual labels:  reinforcement-learning
Mojitalk
Code for "MojiTalk: Generating Emotional Responses at Scale" https://arxiv.org/abs/1711.04090
Stars: ✭ 107 (-6.96%)
Mutual labels:  reinforcement-learning
Handful Of Trials Pytorch
Unofficial Pytorch code for "Deep Reinforcement Learning in a Handful of Trials using Probabilistic Dynamics Models"
Stars: ✭ 112 (-2.61%)
Mutual labels:  reinforcement-learning
Direct Future Prediction Keras
Direct Future Prediction (DFP ) in Keras
Stars: ✭ 103 (-10.43%)
Mutual labels:  reinforcement-learning
Stable Baselines
Mirror of Stable-Baselines: a fork of OpenAI Baselines, implementations of reinforcement learning algorithms
Stars: ✭ 115 (+0%)
Mutual labels:  reinforcement-learning
Numpy Ml
Machine learning, in numpy
Stars: ✭ 11,100 (+9552.17%)
Mutual labels:  reinforcement-learning
Startcraft pysc2 minigames
Startcraft II Machine Learning research with DeepMind pysc2 python library .mini-games and agents.
Stars: ✭ 113 (-1.74%)
Mutual labels:  reinforcement-learning
Easy Rl
强化学习中文教程,在线阅读地址:https://datawhalechina.github.io/easy-rl/
Stars: ✭ 3,004 (+2512.17%)
Mutual labels:  reinforcement-learning
Cartpole
OpenAI's cartpole env solver.
Stars: ✭ 107 (-6.96%)
Mutual labels:  reinforcement-learning
Ctc Executioner
Master Thesis: Limit order placement with Reinforcement Learning
Stars: ✭ 112 (-2.61%)
Mutual labels:  reinforcement-learning
Tensorflow2.0 Examples
🙄 Difficult algorithm, Simple code.
Stars: ✭ 1,397 (+1114.78%)
Mutual labels:  reinforcement-learning
Deep Neuroevolution
Deep Neuroevolution
Stars: ✭ 1,526 (+1226.96%)
Mutual labels:  reinforcement-learning
Reinforcement Learning Cheat Sheet
Reinforcement Learning Cheat Sheet
Stars: ✭ 104 (-9.57%)
Mutual labels:  reinforcement-learning
Navbot
Using RGB Image as Visual Input for Mapless Robot Navigation
Stars: ✭ 111 (-3.48%)
Mutual labels:  reinforcement-learning
Hierarchical Actor Critic Hac Pytorch
PyTorch implementation of Hierarchical Actor Critic (HAC) for OpenAI gym environments
Stars: ✭ 116 (+0.87%)
Mutual labels:  reinforcement-learning
Coursera reinforcement learning
Coursera Reinforcement Learning Specialization by University of Alberta & Alberta Machine Intelligence Institute
Stars: ✭ 114 (-0.87%)
Mutual labels:  reinforcement-learning
Studybook
Study E-Book(ComputerVision DeepLearning MachineLearning Math NLP Python ReinforcementLearning)
Stars: ✭ 1,457 (+1166.96%)
Mutual labels:  reinforcement-learning

Distributional Bellman "C51" Algorithm implemented in Keras

This repo includes implementation of C51 Algorithm describe in this paper. The implementation is tested on the VizDoom Defend the Center scenario, which is a 3D partially observable environment.

For tutorial on Distributional Bellman and step-by-step walkthrough of the implementation, please check out my blog post at https://flyyufelix.github.io/2017/10/24/distributional-bellman.html.

Results

Below is the performance chart of 15,000 episodes of C51 DDQN and DDQN running on Defend the Center. Y-axis is the average number of kills (moving average over 50 episodes).

C51 Performance Chart

Usage

First follow this instruction to install VizDoom. If you use python, you can simply do pip install:

$ pip install vizdoom

Second, clone ViZDoom to your machine, copy the python files provided in this repo over to examples/python.

Next, edit scenarios/defend_the_center.cfg file. Replace this line

available_game_variables = { AMMO2 HEALTH }

with

available_game_variables = { KILLCOUNT AMMO2 HEALTH }

To test if the environment is working, run

$ cd examples/python
$ python c51_ddqn.py

You should see some printouts indicating that the C51 DDQN is running successfully. Errors would be thrown otherwise.

Dependencies

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].