All Projects → jk96491 → SMAC

jk96491 / SMAC

Licence: Apache-2.0 License
StarCraft II Multi Agent Challenge : QMIX, COMA, LIIR, QTRAN, Central V, ROMA, RODE, DOP, Graph MIX

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to SMAC

italy
Free open public domain football data (football.db) for Italy / Europe - Serie A etc.
Stars: ✭ 35 (-12.5%)
Mutual labels:  roma
Meta-Learning-for-StarCraft-II-Minigames
We reproduced DeepMind's results and implement a meta-learning (MLSH) agent which can generalize across minigames.
Stars: ✭ 26 (-35%)
Mutual labels:  starcraft2
sc2reader
Extracts gameplay information from Starcraft II replay files
Stars: ✭ 119 (+197.5%)
Mutual labels:  starcraft2
Auto Sklearn
Automated Machine Learning with scikit-learn
Stars: ✭ 5,916 (+14690%)
Mutual labels:  smac
malib deprecated
A Multi-agent Learning Framework
Stars: ✭ 63 (+57.5%)
Mutual labels:  multi-agent-reinforcement-learning
CommNet-BiCnet
CommNet and BiCnet implementation in tensorflow
Stars: ✭ 50 (+25%)
Mutual labels:  multi-agent-reinforcement-learning
5minBot
Terran Bot for the Sc2 AI ladder at http://sc2ai.net/
Stars: ✭ 27 (-32.5%)
Mutual labels:  starcraft2
mpyq
Python library for reading MPQ archives.
Stars: ✭ 86 (+115%)
Mutual labels:  starcraft2
sc2gym
PySC2 OpenAI Gym Environments
Stars: ✭ 50 (+25%)
Mutual labels:  starcraft2
gym-battlesnake
Multi-agent reinforcement learning environment
Stars: ✭ 29 (-27.5%)
Mutual labels:  multi-agent-reinforcement-learning
Fruit-API
A Universal Deep Reinforcement Learning Framework
Stars: ✭ 61 (+52.5%)
Mutual labels:  multi-agent-reinforcement-learning
HeroesOfTheStorm TryMode2.0
A modified Try Mode Map for Heroes of the Storm to create a better experience with enhanced functionalities.
Stars: ✭ 18 (-55%)
Mutual labels:  starcraft2
cpp-sc2
StarCraft II Client C++ library, proud fork of Blizzard/s2client-api.
Stars: ✭ 31 (-22.5%)
Mutual labels:  starcraft2
pymarl2
Fine-tuned MARL algorithms on SMAC (100% win rates on most scenarios)
Stars: ✭ 311 (+677.5%)
Mutual labels:  smac
StarCraft-Casting-Tool
StarCraft Casting Tool is a free to use open source program that makes casting StarCraft 2 simple while increasing the production value substantially by providing a match grabber and various sets of animated icons and browser sources to be shown to the viewer.
Stars: ✭ 17 (-57.5%)
Mutual labels:  starcraft2
Coach
Reinforcement Learning Coach by Intel AI Lab enables easy experimentation with state of the art Reinforcement Learning algorithms
Stars: ✭ 2,085 (+5112.5%)
Mutual labels:  starcraft2
sc2-planner
No description or website provided.
Stars: ✭ 36 (-10%)
Mutual labels:  starcraft2
Mava
A library of multi-agent reinforcement learning components and systems
Stars: ✭ 355 (+787.5%)
Mutual labels:  multi-agent-reinforcement-learning
CoDAIL
Implementation of CoDAIL in the ICLR 2021 paper <Multi-Agent Interactions Modeling with Correlated Policies>
Stars: ✭ 17 (-57.5%)
Mutual labels:  multi-agent-reinforcement-learning
sharpy-sc2
Python framework for rapid development of Starcraft 2 AI bots
Stars: ✭ 51 (+27.5%)
Mutual labels:  starcraft2

(KOR)StarCraft II Multi Agent Challenge

제공하는 멀티 에이전트 강화학습 모델들 _QMIX, COMA, LIIR, G2ANet, QTRAN, VDN, Central V, IQL, MAVEN, ROMA, RODE, DOP and Graph MIX .

본 저장소는 Windows OS에서 편리하게 실행시키기 위해 만들어짐.

첫번째로 StarCraft 2을 설치 해야 합니다. 체험판도 상관없습니다. 아래 링크에서 받으세요

https://starcraft2.com/ko-kr/

설치 후 미니게임을 위한 맵을 아래 링크에서 다운로드 받아야 합니다.

https://github.com/oxwhirl/smac/tree/master/smac/env/starcraft2/maps/SMAC_Maps

다운 받은 맵들을 아래 경로에 이동시키면 됩니다.

C:\Program Files (x86)\StarCraft II\Maps\SMAC_Maps

이제 부터는 환경 설정 입니다.

필요한 패키지들을 설치하기 위하여 아래대로 명령을 입력하세요

pip install -r requirements.txt

불행이도 한가지는 직접 설치해야 합니다.(어렵지 않습니다.)

pytorch만 아래와 같이 직접 설치해 주세요\

conda install pytorch==1.2.0 torchvision==0.4.0 -c pytorch

마지막으로 "main.py"을 실행 시키면 됩니다.

(ENG)StarCraft II Multi Agent Challenge

The algorithms provided are _QMIX, COMA, LIIR, G2ANet, QTRAN, VDN, Central V, IQL, ROMA, RODE, DOP and Graph MIX .

This repository has been edited for convenient execution in Windows OS.

First you need to install the StarCraft 2 game. Trial version does not matter. Download it from the link below

https://starcraft2.com/ko-kr/

After installation, you should download the map required for the minigame from the link below.

https://github.com/oxwhirl/smac/tree/master/smac/env/starcraft2/maps/SMAC_Maps

You can move all downloaded files to the path below.

C:\Program Files (x86)\StarCraft II\Maps\SMAC_Maps

From now on, this is the environment setting.

Enter the following command to install the packages you need first.

pip install -r requirements.txt

Unfortunately, you need to install the one below yourself (it is not difficult).

You just need to install pytorch.

conda install pytorch==1.2.0 torchvision==0.4.0 -c pytorch

Finally, run "main.py"

Python MARL framework

PyMARL is WhiRL's framework for deep multi-agent reinforcement learning and includes implementations of the following algorithms:

PyMARL is written in PyTorch and uses SMAC as its environment.

Installation instructions

Build the Dockerfile using

cd docker
bash build.sh

Set up StarCraft II and SMAC:

bash install_sc2.sh

This will download SC2 into the 3rdparty folder and copy the maps necessary to run over.

The requirements.txt file can be used to install the necessary packages into a virtual environment (not recomended).

Run an experiment

python3 src/main.py --config=qmix --env-config=sc2 with env_args.map_name=2s3z

The config files act as defaults for an algorithm or environment.

They are all located in src/config. --config refers to the config files in src/config/algs --env-config refers to the config files in src/config/envs

To run experiments using the Docker container:

bash run.sh $GPU python3 src/main.py --config=qmix --env-config=sc2 with env_args.map_name=2s3z

All results will be stored in the Results folder.

The previous config files used for the SMAC Beta have the suffix _beta.

Saving and loading learnt models

Saving models

You can save the learnt models to disk by setting save_model = True, which is set to False by default. The frequency of saving models can be adjusted using save_model_interval configuration. Models will be saved in the result directory, under the folder called models. The directory corresponding each run will contain models saved throughout the experiment, each within a folder corresponding to the number of timesteps passed since starting the learning process.

Loading models

Learnt models can be loaded using the checkpoint_path parameter, after which the learning will proceed from the corresponding timestep.

Watching StarCraft II replays

save_replay option allows saving replays of models which are loaded using checkpoint_path. Once the model is successfully loaded, test_nepisode number of episodes are run on the test mode and a .SC2Replay file is saved in the Replay directory of StarCraft II. Please make sure to use the episode runner if you wish to save a replay, i.e., runner=episode. The name of the saved replay file starts with the given env_args.save_replay_prefix (map_name if empty), followed by the current timestamp.

The saved replays can be watched by double-clicking on them or using the following command:

python -m pysc2.bin.play --norender --rgb_minimap_size 0 --replay NAME.SC2Replay

Note: Replays cannot be watched using the Linux version of StarCraft II. Please use either the Mac or Windows version of the StarCraft II client.

Documentation/Support

Documentation is a little sparse at the moment (but will improve!). Please raise an issue in this repo, or email Tabish

Citing PyMARL

If you use PyMARL in your research, please cite the SMAC paper.

M. Samvelyan, T. Rashid, C. Schroeder de Witt, G. Farquhar, N. Nardelli, T.G.J. Rudner, C.-M. Hung, P.H.S. Torr, J. Foerster, S. Whiteson. The StarCraft Multi-Agent Challenge, CoRR abs/1902.04043, 2019.

In BibTeX format:

@article{samvelyan19smac,
  title = {{The} {StarCraft} {Multi}-{Agent} {Challenge}},
  author = {Mikayel Samvelyan and Tabish Rashid and Christian Schroeder de Witt and Gregory Farquhar and Nantas Nardelli and Tim G. J. Rudner and Chia-Man Hung and Philiph H. S. Torr and Jakob Foerster and Shimon Whiteson},
  journal = {CoRR},
  volume = {abs/1902.04043},
  year = {2019},
}

License

Code licensed under the Apache License v2.0

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].