All Projects → instadeepai → AlphaNPI

instadeepai / AlphaNPI

Licence: MIT license
Adapting the AlphaZero algorithm to remove the need of execution traces to train NPI.

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to AlphaNPI

Carla-ppo
This repository hosts a customized PPO based agent for Carla. The goal of this project is to make it easier to interact with and experiment in Carla with reinforcement learning based agents -- this, by wrapping Carla in a gym like environment that can handle custom reward functions, custom debug output, etc.
Stars: ✭ 122 (+71.83%)
Mutual labels:  deep-reinforcement-learning
DDPG
End to End Mobile Robot Navigation using DDPG (Continuous Control with Deep Reinforcement Learning) based on Tensorflow + Gazebo
Stars: ✭ 41 (-42.25%)
Mutual labels:  deep-reinforcement-learning
rtrl
PyTorch implementation of our paper Real-Time Reinforcement Learning (NeurIPS 2019)
Stars: ✭ 57 (-19.72%)
Mutual labels:  deep-reinforcement-learning
MP-DQN
Source code for the dissertation: "Multi-Pass Deep Q-Networks for Reinforcement Learning with Parameterised Action Spaces"
Stars: ✭ 99 (+39.44%)
Mutual labels:  deep-reinforcement-learning
neural-mpc
No description or website provided.
Stars: ✭ 54 (-23.94%)
Mutual labels:  deep-reinforcement-learning
drl grasping
Deep Reinforcement Learning for Robotic Grasping from Octrees
Stars: ✭ 160 (+125.35%)
Mutual labels:  deep-reinforcement-learning
alpha sigma
A pytorch based Gomoku game model. Alpha Zero algorithm based reinforcement Learning and Monte Carlo Tree Search model.
Stars: ✭ 134 (+88.73%)
Mutual labels:  deep-reinforcement-learning
Underflow
With underflow, create trafic light clusters that interact together to regulate circulation
Stars: ✭ 12 (-83.1%)
Mutual labels:  deep-reinforcement-learning
AI booklet CE-AUT
Booklet and exam of Artificial Intelligence Master Degree at Amirkabir University of technology.
Stars: ✭ 14 (-80.28%)
Mutual labels:  deep-reinforcement-learning
AutoPentest-DRL
AutoPentest-DRL: Automated Penetration Testing Using Deep Reinforcement Learning
Stars: ✭ 196 (+176.06%)
Mutual labels:  deep-reinforcement-learning
CrowdNav DSRNN
[ICRA 2021] Decentralized Structural-RNN for Robot Crowd Navigation with Deep Reinforcement Learning
Stars: ✭ 43 (-39.44%)
Mutual labels:  deep-reinforcement-learning
muzero
A clean implementation of MuZero and AlphaZero following the AlphaZero General framework. Train and Pit both algorithms against each other, and investigate reliability of learned MuZero MDP models.
Stars: ✭ 126 (+77.46%)
Mutual labels:  deep-reinforcement-learning
godpaper
🐵 An AI chess-board-game framework(by many programming languages) implementations.
Stars: ✭ 40 (-43.66%)
Mutual labels:  deep-reinforcement-learning
code summarization public
source code for 'Improving automatic source code summarization via deep reinforcement learning'
Stars: ✭ 71 (+0%)
Mutual labels:  deep-reinforcement-learning
racing dreamer
Latent Imagination Facilitates Zero-Shot Transfer in Autonomous Racing
Stars: ✭ 31 (-56.34%)
Mutual labels:  deep-reinforcement-learning
DeepLearningFlappyFrog
Flappy Frog hack using Deep Reinforcement Learning (Deep Q-learning). 暴力膜蛤不可取。
Stars: ✭ 16 (-77.46%)
Mutual labels:  deep-reinforcement-learning
datascience-mashup
In this repo I will try to gather all of the projects related to data science with clean datasets and high accuracy models to solve real world problems.
Stars: ✭ 36 (-49.3%)
Mutual labels:  deep-reinforcement-learning
Explorer
Explorer is a PyTorch reinforcement learning framework for exploring new ideas.
Stars: ✭ 54 (-23.94%)
Mutual labels:  deep-reinforcement-learning
semantic-guidance
Code for our CVPR-2021 paper on Combining Semantic Guidance and Deep Reinforcement Learning For Generating Human Level Paintings.
Stars: ✭ 19 (-73.24%)
Mutual labels:  deep-reinforcement-learning
pokeai
Develop ultimate AI Pokémon trainer
Stars: ✭ 18 (-74.65%)
Mutual labels:  deep-reinforcement-learning

AlphaNPI

Adapting the AlphaZero algorithm to remove the need of execution traces to train NPI.

Setup

You need to install the required Python packages.

cd alphanpi/

Then run the command:

pip install -r requirements.txt

Update the following environment variable:

export PYTHONPATH=$PWD:$PYTHONPATH

Training

cd trainings/

Run one of the scripts:

python train_recursive_sorting.py --tensorboard --verbose --save-model --save-results --save-model
python train_hanoi.py --tensorboard --verbose --save-model --save-results --save-model
python train_sorting_nohierarchy.py --tensorboard --verbose --save-model --save-results --save-model
python train_hanoi.py --tensorboard --verbose --save-model --save-results --save-model

For more information about the arguments that can be sent, see at trainings/README.md

Validation

The following allows to assert the results disclosed in the paper

cd validation/

Then run one of the scripts:

python validate_hanoi.py --verbose --save-results
python validate_recursive_sorting.py --verbose --save-results
python validate_sorting.py --verbose --save-results
python validate_sorting_nohierarchy.py --verbose --save-results

For more information about the arguments that can be sent, see at validation/README.md

Visualization

cd visualization/

To visualize a pre-trained model behavior for an environment env, run the script visualize_{env}.py. Set the load path at the beginning of the script to the path where the model of interest is saved.

Run one of the scripts:

python visualize_hanoi.py
python visualize_recursive_sorting.py
python visualize_sorting.py
python visualize_sorting_nohierarchy.py

Generate visualization

When a visualization script is executed, it generates a mcts.gv file under visualization/. The file contains a description of the tree in dot language.

If you don't already have graphviz installed, run command:

sudo apt-get install python3-pydot graphviz

To convert the .gv in .pdf file, use command:

dot -Tpdf mcts.gv -o mcts.pdf
Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].