All Projects → Unity-Technologies → Obstacle Tower Env

Unity-Technologies / Obstacle Tower Env

Licence: apache-2.0
Obstacle Tower Environment

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to Obstacle Tower Env

Smac
SMAC: The StarCraft Multi-Agent Challenge
Stars: ✭ 435 (-12.3%)
Mutual labels:  reinforcement-learning
Recsim
A Configurable Recommender Systems Simulation Platform
Stars: ✭ 461 (-7.06%)
Mutual labels:  reinforcement-learning
Awesome Robotics
A curated list of awesome links and software libraries that are useful for robots.
Stars: ✭ 478 (-3.63%)
Mutual labels:  reinforcement-learning
Mushroom Rl
Python library for Reinforcement Learning.
Stars: ✭ 442 (-10.89%)
Mutual labels:  reinforcement-learning
Courses
Quiz & Assignment of Coursera
Stars: ✭ 454 (-8.47%)
Mutual labels:  reinforcement-learning
Robosuite
robosuite: A Modular Simulation Framework and Benchmark for Robot Learning
Stars: ✭ 462 (-6.85%)
Mutual labels:  reinforcement-learning
Autonomous Learning Library
A PyTorch library for building deep reinforcement learning agents.
Stars: ✭ 425 (-14.31%)
Mutual labels:  reinforcement-learning
Ml Mipt
Open Machine Learning course at MIPT
Stars: ✭ 480 (-3.23%)
Mutual labels:  reinforcement-learning
Arnold
Arnold - DOOM Agent
Stars: ✭ 457 (-7.86%)
Mutual labels:  reinforcement-learning
Tensorflow Reinforce
Implementations of Reinforcement Learning Models in Tensorflow
Stars: ✭ 480 (-3.23%)
Mutual labels:  reinforcement-learning
Tensor House
A collection of reference machine learning and optimization models for enterprise operations: marketing, pricing, supply chain
Stars: ✭ 449 (-9.48%)
Mutual labels:  reinforcement-learning
Robotics Rl Srl
S-RL Toolbox: Reinforcement Learning (RL) and State Representation Learning (SRL) for Robotics
Stars: ✭ 453 (-8.67%)
Mutual labels:  reinforcement-learning
Neurojs
A JavaScript deep learning and reinforcement learning library.
Stars: ✭ 4,344 (+775.81%)
Mutual labels:  reinforcement-learning
Reinforcement learning tutorial with demo
Reinforcement Learning Tutorial with Demo: DP (Policy and Value Iteration), Monte Carlo, TD Learning (SARSA, QLearning), Function Approximation, Policy Gradient, DQN, Imitation, Meta Learning, Papers, Courses, etc..
Stars: ✭ 442 (-10.89%)
Mutual labels:  reinforcement-learning
Rl a3c pytorch
A3C LSTM Atari with Pytorch plus A3G design
Stars: ✭ 482 (-2.82%)
Mutual labels:  reinforcement-learning
Qlearning trading
Learning to trade under the reinforcement learning framework
Stars: ✭ 431 (-13.1%)
Mutual labels:  reinforcement-learning
Rl Book
Source codes for the book "Reinforcement Learning: Theory and Python Implementation"
Stars: ✭ 464 (-6.45%)
Mutual labels:  reinforcement-learning
Autokernel
AutoKernel 是一个简单易用,低门槛的自动算子优化工具,提高深度学习算法部署效率。
Stars: ✭ 485 (-2.22%)
Mutual labels:  reinforcement-learning
Tensorflow Book
Accompanying source code for Machine Learning with TensorFlow. Refer to the book for step-by-step explanations.
Stars: ✭ 4,448 (+796.77%)
Mutual labels:  reinforcement-learning
Practical rl
A course in reinforcement learning in the wild
Stars: ✭ 4,741 (+855.85%)
Mutual labels:  reinforcement-learning

Obstacle Tower Environment

alt text

About

The Obstacle Tower is a procedurally generated environment consisting of multiple floors to be solved by a learning agent. It is designed to test learning agents abilities in computer vision, locomotion skills, high-level planning, and generalization. It combines platforming-style gameplay with puzzles and planning problems, and critically, increases in difficulty as the agent progresses.

Within each floor, the goal of the agent is to arrive at the set of stairs leading to the next level of the tower. These floors are composed of multiple rooms, each which can contain their own unique challenges. Furthermore, each floor contains a number of procedurally generated elements, such as visual appearance, puzzle configuration, and floor layout. This ensures that in order for an agent to be successful at the Obstacle Tower task, they must be able to generalize to new and unseen combinations of conditions.

Reference Paper

To learn more, please read our AAAI Workshop paper:

Obstacle Tower: A Generalization Challenge in Vision, Control, and Planning.

Version History

  • v1.0 - Initial Release.
  • v1.1 - Obstacle Tower Challenge Round 1 Release.
    • Improved determinism between resets.
    • Fixed bug in room template leading to un-beatable floors.
    • Improved rendering & communication speed.
  • v1.2 - Hotfix release.
    • Adds timeout_wait parameter to extend python wait time for Unity environment handshake.
    • Adds realtime_mode parameter to launch Unity environment from API at real-time speed and render to the window.
    • Updates Windows and Linux binaries to address launching issues.
    • Updated v1.2 binary includes fixes for agent collision detection issues.
  • v1.3 Hotfix release.
    • Resolves memory leak when running in Docker.
    • Fixes issue where environment could freeze on certain higher floors.
  • v2.0 Obstacle Tower Challenge Round 2 Release.
    • Towers can now be generated with up to 100 floors.
    • Additional visual themes, obstacles, enemy types, and floor difficulties added.
    • Additional reset parameters added to customize generated towers. Go here for details on the parameters and their values.
    • Various bugs fixed and performance improvements.
  • v2.1 Hotfix release.
    • Resolves issue with new seeds not being applied on env.reset.
    • Resolves issue with underspecified observation space.
  • v2.2 Hotfix release.
    • Resolves issue with reset parameters sometimes not being updated during env.reset.
    • Resolves issue where agents could possibly skip levels.
  • v3.0 Open Source release available here.
    • Added evaluation script for benchmarking, along with official benchmarking guidelines.
  • v3.1 Minor release.
    • Updates Unity project to 2019.2 and ML-Agents version to 0.10.
  • 4.0 Major project updates
    • Unity project updated to 2019.4 LTS.
    • ML-Agents package updated to 1.1.0 (Release 3).
    • Gym wrapper now uses registry by default to auto-download binary.
  • 4.1 Minor release.
    • Adds additional reset parameters for more fine-grained control of visual themes.
    • Fixes issue in gym wrapper related to end of episode information.
    • Fixes issue with environment not being responsive to some reset parameters.

Installation

Requirements

The Obstacle Tower environment runs on Mac OS X, Windows, or Linux.

Python dependencies (also in setup.py):

  • Python 3.6+
  • Unity ML-Agents 1.x
  • OpenAI Gym

Download the environment (optional)

By default, the binary will be automatically downloaded when the Obstacle Tower gym is first instantiated. The binaries for each platform can be separately downloaded at the following links.

Platform Download Link
Linux (x86_64) https://storage.googleapis.com/obstacle-tower-build/v4.1/obstacletower_v4.1_linux.zip
Mac OS X https://storage.googleapis.com/obstacle-tower-build/v4.1/obstacletower_v4.1_osx.zip
Windows https://storage.googleapis.com/obstacle-tower-build/v4.1/obstacletower_v4.1_windows.zip

For checksums on these files, see here.

Install the Gym interface

$ git clone [email protected]:Unity-Technologies/obstacle-tower-env.git
$ cd obstacle-tower-env
$ pip install -e .

Getting Started

Using the Gym Interface

To see an example of how to interact with the environment using the gym interface, see our Basic Usage Jupyter Notebook.

Customizing the environment

Obstacle Tower can be configured in a number of different ways to adjust the difficulty and content of the environment. This is done through the use of reset parameters, which can be set when calling env.reset(). See here for a list of the available parameters to adjust.

Player Control

It is also possible to launch the environment in "Player Mode," and directly control the agent using a keyboard. This can be done by double-clicking on the binary file. The keyboard controls are as follows:

Keyboard Key Action
W Move character forward.
S Move character backwards.
A Move character left.
D Move character right.
K Rotate camera left.
L Rotate camera right.
Space Character jump.

Performance evaluation

We provide an environment wrapper for evaluating performance of a player or agent across multiple pre-defined seeds. We provide an example implementation of evaluation on a random policy.

Training a Dopamine Rainbow agent on GCP

If you are interested in training an agent using Google's Dopamine framework and/or Google Cloud Platform, see our guide here.

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].