All Projects → tegg89 → Magnet

tegg89 / Magnet

Licence: apache-2.0
MAGNet: Multi-agents control using Graph Neural Networks

Projects that are alternatives of or similar to Magnet

Awesome Ai Books
Some awesome AI related books and pdfs for learning and downloading, also apply some playground models for learning
Stars: ✭ 855 (+871.59%)
Mutual labels:  jupyter-notebook, reinforcement-learning
Policy Gradient Methods
Implementation of Algorithms from the Policy Gradient Family. Currently includes: A2C, A3C, DDPG, TD3, SAC
Stars: ✭ 54 (-38.64%)
Mutual labels:  jupyter-notebook, reinforcement-learning
Udacity Deep Learning Nanodegree
This is just a collection of projects that made during my DEEPLEARNING NANODEGREE by UDACITY
Stars: ✭ 15 (-82.95%)
Mutual labels:  jupyter-notebook, reinforcement-learning
Basic reinforcement learning
An introductory series to Reinforcement Learning (RL) with comprehensive step-by-step tutorials.
Stars: ✭ 826 (+838.64%)
Mutual labels:  jupyter-notebook, reinforcement-learning
Rl Course Experiments
Stars: ✭ 73 (-17.05%)
Mutual labels:  jupyter-notebook, reinforcement-learning
Deeplearning Trader
backtrader with DRL ( Deep Reinforcement Learning)
Stars: ✭ 24 (-72.73%)
Mutual labels:  jupyter-notebook, reinforcement-learning
Machine Learning From Scratch
Succinct Machine Learning algorithm implementations from scratch in Python, solving real-world problems (Notebooks and Book). Examples of Logistic Regression, Linear Regression, Decision Trees, K-means clustering, Sentiment Analysis, Recommender Systems, Neural Networks and Reinforcement Learning.
Stars: ✭ 42 (-52.27%)
Mutual labels:  jupyter-notebook, reinforcement-learning
Reinforcement Learning 2nd Edition By Sutton Exercise Solutions
Solutions of Reinforcement Learning, An Introduction
Stars: ✭ 713 (+710.23%)
Mutual labels:  jupyter-notebook, reinforcement-learning
Rl Workshop
Reinforcement Learning Workshop for Data Science BKK
Stars: ✭ 73 (-17.05%)
Mutual labels:  jupyter-notebook, reinforcement-learning
Reinforcement Learning
Implementation of Reinforcement Learning algorithms in Python, based on Sutton's & Barto's Book (Ed. 2)
Stars: ✭ 55 (-37.5%)
Mutual labels:  jupyter-notebook, reinforcement-learning
Coursera
Quiz & Assignment of Coursera
Stars: ✭ 774 (+779.55%)
Mutual labels:  jupyter-notebook, reinforcement-learning
Mathy
Tools for using computer algebra systems to solve math problems step-by-step with reinforcement learning
Stars: ✭ 79 (-10.23%)
Mutual labels:  jupyter-notebook, reinforcement-learning
Hands On Meta Learning With Python
Learning to Learn using One-Shot Learning, MAML, Reptile, Meta-SGD and more with Tensorflow
Stars: ✭ 768 (+772.73%)
Mutual labels:  jupyter-notebook, reinforcement-learning
Rainbow Is All You Need
Rainbow is all you need! A step-by-step tutorial from DQN to Rainbow
Stars: ✭ 938 (+965.91%)
Mutual labels:  jupyter-notebook, reinforcement-learning
Deeprl Tutorials
Contains high quality implementations of Deep Reinforcement Learning algorithms written in PyTorch
Stars: ✭ 748 (+750%)
Mutual labels:  jupyter-notebook, reinforcement-learning
World Models Sonic Pytorch
Attempt at reinforcement learning with curiosity for Sonic the Hedgehog games. Number 149 on OpenAI retro contest leaderboard, but more work needed
Stars: ✭ 27 (-69.32%)
Mutual labels:  jupyter-notebook, reinforcement-learning
David Silver Reinforcement Learning
Notes for the Reinforcement Learning course by David Silver along with implementation of various algorithms.
Stars: ✭ 623 (+607.95%)
Mutual labels:  jupyter-notebook, reinforcement-learning
Hands On Reinforcement Learning With Python
Master Reinforcement and Deep Reinforcement Learning using OpenAI Gym and TensorFlow
Stars: ✭ 640 (+627.27%)
Mutual labels:  jupyter-notebook, reinforcement-learning
Notebooks
Some notebooks
Stars: ✭ 53 (-39.77%)
Mutual labels:  jupyter-notebook, reinforcement-learning
Reinforcement Learning
Reinforcement learning material, code and exercises for Udacity Nanodegree programs.
Stars: ✭ 77 (-12.5%)
Mutual labels:  jupyter-notebook, reinforcement-learning

Deep Multi-Agent Reinforcement Learning with Relevance Graphs

Code for the Deep Multi-Agent Reinforcement Learning with Relevance Graphs accepted from NeurIPS Deep RL workshop.

Objective

The goal of this project is controlling multi-agents using reinforcement learning and graph neural networks. Multi-agent scenarios are usually sparsely rewarded. Graph neural networks have an advantage that each node can be trained robustly. With this property, we hypothesized that each agent in an environment can be controlled individually. Since there have been many research papers related to graph neural networks, we would like to apply it to reinforcement learning.

For the experiment, we will use Pommerman environment. This has relatively strict constraints on environment settings and simple to deploy algorithms.

Proposed methods

  • The proposed architectures is structured with two stages, graph construction and optimal action execution.
  • Inspired by the curiosity-driven paper, we use self-supervised prediction to infer environments, constructing graph matrix. Taking the concatenation of previous states and actions, the graph is constructed. This stage is solving a regression problem in supervised learning.
  • Afterward, the trained graph goes through NerveNet to perform an action. Also, the graph goes to MLP with concatenated state and action value to produce action value. Those two values are compared and trained using DDPG algorithm.
  • The design of the network is shown below.

Dependencies

The script has been tested running under Python 3.6.6, with the following packages installed (along with their dependencies):

  • numpy==1.14.5
  • tensorflow==1.8.0

Experiments

1. Environements

  • Pommerman is sponsored by NVIDIA, FAIR, and Google AI.
  • For each agent: 372 observation spaces (board, bomb_blast strength, bomb_life, position, blast strength, can kick, teammate, ammo, enemies) & 6 action spaces (up, down, right, left, bomb, stop)
  • Free for all & Team match modes are available.

2. Results

2-1. Algorithm comparisons

  • Top graph shows the performance of the proposed model and other RL algorithms.
  • Bottom graph shows the effectiveness of graph construction.
  • Graph sharing within the team and individual graphs per agents are tested, and with the shared graph construction model gains the better performance.

2-2. Graph evaluations

  • We experimented with an effectiveness of shared and individual graphs. We set individual graphs per agents to the opposite side and take the shared graph to the allied side. As we training the model, the shared graph has better performance over separated ones.

  • The visualizations of constructed graphs are shown in this result. The left graph shows the shared graph, whereas the right graph shows the separated graphs.
  • At the beginning of the game (top), the graph is ordered with same structures with equally distributed edges.
  • In the middle of the game (bottom), shared graph one shows the same team chases one of the opponent agents. In the separated graph, all agents are evenly chasing to each other.

Authors

Tegg Taekyong Sung & Aleksandra Malysheva

Acknowledgement

This was supported by Deep Learning Camp Jeju 2018 which was organized by TensorFlow Korea User Group.

License

Apache

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].