All Projects → higgsfield → Rl Adventure

higgsfield / Rl Adventure

Pytorch Implementation of DQN / DDQN / Prioritized replay/ noisy networks/ distributional values/ Rainbow/ hierarchical RL

Programming Languages

Jupyter Notebook
11667 projects
python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to Rl Adventure

Up Down Captioner
Automatic image captioning model based on Caffe, using features from bottom-up attention.
Stars: ✭ 195 (-92.22%)
Mutual labels:  jupyter-notebook
Nas fpn tensorflow
NAS-FPN: Learning Scalable Feature Pyramid Architecture for Object Detection.
Stars: ✭ 198 (-92.1%)
Mutual labels:  jupyter-notebook
Mgcnn
Multi-Graph Convolutional Neural Networks
Stars: ✭ 199 (-92.06%)
Mutual labels:  jupyter-notebook
Auto Reid And Others
Auto-ReID and Other Person Re-Identification Projects
Stars: ✭ 198 (-92.1%)
Mutual labels:  jupyter-notebook
Neural networks and cv
Stars: ✭ 199 (-92.06%)
Mutual labels:  jupyter-notebook
Go Tflite
Go binding for TensorFlow Lite
Stars: ✭ 199 (-92.06%)
Mutual labels:  jupyter-notebook
Data Science Projects With Python
A Case Study Approach to Successful Data Science Projects Using Python, Pandas, and Scikit-Learn
Stars: ✭ 198 (-92.1%)
Mutual labels:  jupyter-notebook
Food2vec
🍔
Stars: ✭ 199 (-92.06%)
Mutual labels:  jupyter-notebook
Pytorch Geometric Yoochoose
This is a tutorial for PyTorch Geometric on the YooChoose dataset
Stars: ✭ 198 (-92.1%)
Mutual labels:  jupyter-notebook
Radio
RadIO is a library for data science research of computed tomography imaging
Stars: ✭ 198 (-92.1%)
Mutual labels:  jupyter-notebook
Bilateral solver
Python code for the fast bilateral solver
Stars: ✭ 198 (-92.1%)
Mutual labels:  jupyter-notebook
Graph Notebook
Library extending Jupyter notebooks to integrate with Apache TinkerPop and RDF SPARQL.
Stars: ✭ 199 (-92.06%)
Mutual labels:  jupyter-notebook
Ml Lessons
Intro to deep learning for medical imaging lesson, by MD.ai
Stars: ✭ 199 (-92.06%)
Mutual labels:  jupyter-notebook
Integratedgradients
Python/Keras implementation of integrated gradients presented in "Axiomatic Attribution for Deep Networks" for explaining any model defined in Keras framework.
Stars: ✭ 197 (-92.14%)
Mutual labels:  jupyter-notebook
Datascience
책) 파이썬으로 데이터 주무르기 - 소스코드 및 데이터 공개
Stars: ✭ 199 (-92.06%)
Mutual labels:  jupyter-notebook
Atari Model Zoo
A binary release of trained deep reinforcement learning models trained in the Atari machine learning benchmark, and a software release that enables easy visualization and analysis of models, and comparison across training algorithms.
Stars: ✭ 198 (-92.1%)
Mutual labels:  jupyter-notebook
Python For Finance Cookbook
Python for Finance Cookbook, published by Packt
Stars: ✭ 199 (-92.06%)
Mutual labels:  jupyter-notebook
Neuralnetworks.thought Experiments
Observations and notes to understand the workings of neural network models and other thought experiments using Tensorflow
Stars: ✭ 199 (-92.06%)
Mutual labels:  jupyter-notebook
Pysonar
Decentralized Machine Learning Client
Stars: ✭ 199 (-92.06%)
Mutual labels:  jupyter-notebook
Trading Gym
This trading-gym is the first trading for agent to train with episode of short term trading itself.
Stars: ✭ 194 (-92.26%)
Mutual labels:  jupyter-notebook

DQN Adventure: from Zero to State of the Art

This is easy-to-follow step-by-step Deep Q Learning tutorial with clean readable code.

The deep reinforcement learning community has made several independent improvements to the DQN algorithm. This tutorial presents latest extensions to the DQN algorithm in the following order:

  1. Playing Atari with Deep Reinforcement Learning [arxiv] [code]
  2. Deep Reinforcement Learning with Double Q-learning [arxiv] [code]
  3. Dueling Network Architectures for Deep Reinforcement Learning [arxiv] [code]
  4. Prioritized Experience Replay [arxiv] [code]
  5. Noisy Networks for Exploration [arxiv] [code]
  6. A Distributional Perspective on Reinforcement Learning [arxiv] [code]
  7. Rainbow: Combining Improvements in Deep Reinforcement Learning [arxiv] [code]
  8. Distributional Reinforcement Learning with Quantile Regression [arxiv] [code]
  9. Hierarchical Deep Reinforcement Learning: Integrating Temporal Abstraction and Intrinsic Motivation [arxiv] [code]
  10. Neural Episodic Control [arxiv] [code]

Environments

First, I recommend to use small test problems to run experiments quickly. Then, you can continue on environments with large observation space.

  • CartPole - classic RL environment can be solved on a single cpu
  • Atari Pong - the easiest atari environment, only takes ~ 1 million frames to converge, comparing with other atari games that take > 40 millions
  • Atari others - change hyperparameters, target network update frequency=10K, replay buffer size=1M

If you get stuck…

  • Remember you are not stuck unless you have spent more than a week on a single algorithm. It is perfectly normal if you do not have all the required knowledge of mathematics and CS. For example, you will need knowledge of the fundamentals of measure theory and statistics, especially the Wasserstein metric and quantile regression. Statistical inference: importance sampling. Data structures: Segment Tree and K-dimensional Tree.
  • Carefully go through the paper. Try to see what is the problem the authors are solving. Understand a high-level idea of the approach, then read the code (skipping the proofs), and after go over the mathematical details and proofs.

Best RL courses

  • David Silver's course link
  • Berkeley deep RL link
  • Practical RL link
Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].