All Projects → CR-Gjx → Leakgan

CR-Gjx / Leakgan

The codes of paper "Long Text Generation via Adversarial Training with Leaked Information" on AAAI 2018. Text generation using GAN and Hierarchical Reinforcement Learning.

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to Leakgan

Awesome Tensorlayer
A curated list of dedicated resources and applications
Stars: ✭ 248 (-53.47%)
Mutual labels:  natural-language-processing, reinforcement-learning, generative-adversarial-network
Conversational Ai
Conversational AI Reading Materials
Stars: ✭ 34 (-93.62%)
Mutual labels:  natural-language-processing, reinforcement-learning, generative-adversarial-network
Textgan Pytorch
TextGAN is a PyTorch framework for Generative Adversarial Networks (GANs) based text generation models.
Stars: ✭ 479 (-10.13%)
Mutual labels:  natural-language-processing, generative-adversarial-network, text-generation
Matterport3dsimulator
AI Research Platform for Reinforcement Learning from Real Panoramic Images.
Stars: ✭ 260 (-51.22%)
Mutual labels:  natural-language-processing, reinforcement-learning
Summarization Papers
Summarization Papers
Stars: ✭ 238 (-55.35%)
Mutual labels:  natural-language-processing, text-generation
skip-thought-gan
Generating Text through Adversarial Training(GAN) using Skip-Thought Vectors
Stars: ✭ 44 (-91.74%)
Mutual labels:  text-generation, generative-adversarial-network
Minerva
Meandering In Networks of Entities to Reach Verisimilar Answers
Stars: ✭ 205 (-61.54%)
Mutual labels:  natural-language-processing, reinforcement-learning
Awesome Self Supervised Learning
A curated list of awesome self-supervised methods
Stars: ✭ 4,492 (+742.78%)
Mutual labels:  natural-language-processing, reinforcement-learning
Deep Learning Nlp Rl Papers
Recent Deep Learning papers in NLU and RL
Stars: ✭ 288 (-45.97%)
Mutual labels:  natural-language-processing, reinforcement-learning
Tensorlayer Tricks
How to use TensorLayer
Stars: ✭ 357 (-33.02%)
Mutual labels:  natural-language-processing, reinforcement-learning
Mlinterview
A curated awesome list of AI Startups in India & Machine Learning Interview Guide. Feel free to contribute!
Stars: ✭ 410 (-23.08%)
Mutual labels:  natural-language-processing, reinforcement-learning
Catalyst
Accelerated deep learning R&D
Stars: ✭ 2,804 (+426.08%)
Mutual labels:  natural-language-processing, reinforcement-learning
Pytorch Rl
This repository contains model-free deep reinforcement learning algorithms implemented in Pytorch
Stars: ✭ 394 (-26.08%)
Mutual labels:  reinforcement-learning, generative-adversarial-network
Courses
Quiz & Assignment of Coursera
Stars: ✭ 454 (-14.82%)
Mutual labels:  natural-language-processing, reinforcement-learning
Textbox
TextBox is an open-source library for building text generation system.
Stars: ✭ 257 (-51.78%)
Mutual labels:  generative-adversarial-network, text-generation
Machine Learning Notebooks
Machine Learning notebooks for refreshing concepts.
Stars: ✭ 222 (-58.35%)
Mutual labels:  natural-language-processing, reinforcement-learning
Arxivtimes
repository to research & share the machine learning articles
Stars: ✭ 3,651 (+584.99%)
Mutual labels:  natural-language-processing, reinforcement-learning
Deep Generative Models For Natural Language Processing
DGMs for NLP. A roadmap.
Stars: ✭ 185 (-65.29%)
Mutual labels:  natural-language-processing, text-generation
Attention Mechanisms
Implementations for a family of attention mechanisms, suitable for all kinds of natural language processing tasks and compatible with TensorFlow 2.0 and Keras.
Stars: ✭ 203 (-61.91%)
Mutual labels:  natural-language-processing, text-generation
Awesome Text Generation
A curated list of recent models of text generation and application
Stars: ✭ 370 (-30.58%)
Mutual labels:  natural-language-processing, text-generation

LeakGAN

The code of research paper Long Text Generation via Adversarial Training with Leaked Information.

This paper has been accepted at the Thirty-Second AAAI Conference on Artificial Intelligence (AAAI-18).

Requirements

  • Tensorflow r1.2.1
  • Python 2.7
  • CUDA 7.5+ (For GPU)

Introduction

Automatically generating coherent and semantically meaningful text has many applications in machine translation, dialogue systems, image captioning, etc. Recently, by combining with policy gradient, Generative Adversarial Nets (GAN) that use a discriminative model to guide the training of the generative model as a reinforcement learning policy has shown promising results in text generation. However, the scalar guiding signal is only available after the entire text has been generated and lacks intermediate information about text structure during the generative process. As such, it limits its success when the length of the generated text samples is long (more than 20 words). In this project, we propose a new framework, called LeakGAN, to address the problem for long text generation. We allow the discriminative net to leak its own high-level extracted features to the generative net to further help the guidance. The generator incorporates such informative signals into all generation steps through an additional Manager module, which takes the extracted features of current generated words and outputs a latent vector to guide the Worker module for next-word generation. Our extensive experiments on synthetic data and various real-world tasks with Turing test demonstrate that LeakGAN is highly effective in long text generation and also improves the performance in short text generation scenarios. More importantly, without any supervision, LeakGAN would be able to implicitly learn sentence structures only through the interaction between Manager and Worker.

As the illustration of LeakGAN. We specifically introduce a hierarchical generator G, which consists of a high-level MANAGER module and a low-level WORKER module. The MANAGER is a long short term memory network (LSTM) and serves as a mediator. In each step, it receives generator D’s high-level feature representation, e.g., the feature map of the CNN, and uses it to form the guiding goal for the WORKER module in that timestep. As the information from D is internally-maintained and in an adversarial game it is not supposed to provide G with such information. We thus call it a leakage of information from D.

Next, given the goal embedding produced by the MANAGER, the WORKER firstly encodes current generated words with another LSTM, then combines the output of the LSTM and the goal embedding to take a final action at current state. As such, the guiding signals from D are not only available to G at the end in terms of the scalar reward signals, but also available in terms of a goal embedding vector during the generation process to guide G how to get improved.

Reference

@article{guo2017long,
  title={Long Text Generation via Adversarial Training with Leaked Information},
  author={Guo, Jiaxian and Lu, Sidi and Cai, Han and Zhang, Weinan and Yu, Yong and Wang, Jun},
  journal={arXiv preprint arXiv:1709.08624},
  year={2017}
}

You can get the code and run the experiments in follow folders.

Folder

Synthetic Data: synthetic data experiment

Image COCO: a real text example for our model using dataset Image COCO (http://cocodataset.org/#download)

Note: this code is based on the previous work by LantaoYu. Many thanks to LantaoYu.

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].