All Projects → morningmoni → Taxorl

morningmoni / Taxorl

Licence: mit
Code for paper "End-to-End Reinforcement Learning for Automatic Taxonomy Induction", ACL 2018

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to Taxorl

Biped trajectory optimization
Implementing trajectory optimization on bipedal system
Stars: ✭ 45 (-21.05%)
Mutual labels:  reinforcement-learning
Gym Minigrid
Minimalistic gridworld package for OpenAI Gym
Stars: ✭ 1,047 (+1736.84%)
Mutual labels:  reinforcement-learning
Reinforcepy
Collection of reinforcement learners implemented in python. Mainly including DQN and its variants
Stars: ✭ 54 (-5.26%)
Mutual labels:  reinforcement-learning
Deep traffic
MIT DeepTraffic top 2% solution (75.01 mph) 🚗.
Stars: ✭ 47 (-17.54%)
Mutual labels:  reinforcement-learning
Gbrain
GPU Javascript Library for Machine Learning
Stars: ✭ 48 (-15.79%)
Mutual labels:  reinforcement-learning
Policy Gradient Methods
Implementation of Algorithms from the Policy Gradient Family. Currently includes: A2C, A3C, DDPG, TD3, SAC
Stars: ✭ 54 (-5.26%)
Mutual labels:  reinforcement-learning
Deterministic Gail Pytorch
PyTorch implementation of Deterministic Generative Adversarial Imitation Learning (GAIL) for Off Policy learning
Stars: ✭ 44 (-22.81%)
Mutual labels:  reinforcement-learning
Ml Surveys
📋 Survey papers summarizing advances in deep learning, NLP, CV, graphs, reinforcement learning, recommendations, graphs, etc.
Stars: ✭ 1,063 (+1764.91%)
Mutual labels:  reinforcement-learning
Holodeck Engine
High Fidelity Simulator for Reinforcement Learning and Robotics Research.
Stars: ✭ 48 (-15.79%)
Mutual labels:  reinforcement-learning
Dqn
Implementation of q-learning using TensorFlow
Stars: ✭ 53 (-7.02%)
Mutual labels:  reinforcement-learning
Reinforcement learning
Predict/log/learn/update reinforcement learning loop
Stars: ✭ 47 (-17.54%)
Mutual labels:  reinforcement-learning
Dher
DHER: Hindsight Experience Replay for Dynamic Goals (ICLR-2019)
Stars: ✭ 48 (-15.79%)
Mutual labels:  reinforcement-learning
Notebooks
Some notebooks
Stars: ✭ 53 (-7.02%)
Mutual labels:  reinforcement-learning
Ml In Tf
Get started with Machine Learning in TensorFlow with a selection of good reads and implemented examples!
Stars: ✭ 45 (-21.05%)
Mutual labels:  reinforcement-learning
Demos
Some JavaScript works published as demos, mostly ML or DS
Stars: ✭ 55 (-3.51%)
Mutual labels:  reinforcement-learning
Async Deeprl
Playing Atari games with TensorFlow implementation of Asynchronous Deep Q-Learning
Stars: ✭ 44 (-22.81%)
Mutual labels:  reinforcement-learning
Pytorch Rl
Stars: ✭ 52 (-8.77%)
Mutual labels:  reinforcement-learning
Tictactoe
Tic Tac Toe Machine Learning
Stars: ✭ 56 (-1.75%)
Mutual labels:  reinforcement-learning
Reinforcement Learning
Implementation of Reinforcement Learning algorithms in Python, based on Sutton's & Barto's Book (Ed. 2)
Stars: ✭ 55 (-3.51%)
Mutual labels:  reinforcement-learning
Notes
The notes for Math, Machine Learning, Deep Learning and Research papers.
Stars: ✭ 53 (-7.02%)
Mutual labels:  reinforcement-learning

TaxoRL

code for "End-to-End Reinforcement Learning for Automatic Taxonomy Induction" ACL 2018 [arXiv]

Requirements

python 2.7
dynet 2.0
tqdm

Data

Preprocessed pickled data including everything else for the WordNet data can be downloaded here.

Preprocessed pickled data including everything else for the WordNet data and SemEval-2016 can be downloaded here. If you run on SemEval-2016, use dev_twodatasets.tsv instead of dev_wnbo_hyper.tsv. Caution: it may take 40+ GB memory.

Go to https://morningmoni.github.io/wordnet-vis/ to see the visualization of WordNet subtrees.

DIY

  • To build everything from scratch, first download corpora such as Wikipedia, UMBC, and 1 Billion Word Language Model Benchmark.
  • To preprocess the corpus, generate a vocabulary file and use the scripts under ./corpus/ to find dependency paths between terms in the vocabulary. The scripts are modified based on LexNET. Instructions can be found here. It may take several hours to finish this process.
  • Run train_RL.py and it will compute all the features and save them into pickle files.

Run

Run train_RL.py for training and testing. All the parameters are in argparse and have default values so that you can run without specifying any parameters (but feel free to tune them).

In each epoch, the performance on training/validation/test sets is reported. You may exit the program at any time.

Cite

@InProceedings{P18-1229,
  author = 	"Mao, Yuning
		and Ren, Xiang
		and Shen, Jiaming
		and Gu, Xiaotao
		and Han, Jiawei",
  title = 	"End-to-End Reinforcement Learning for Automatic Taxonomy Induction",
  booktitle = 	"Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)",
  year = 	"2018",
  publisher = 	"Association for Computational Linguistics",
  pages = 	"2462--2472",
  location = 	"Melbourne, Australia",
  url = 	"http://aclweb.org/anthology/P18-1229"
}
Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].