All Projects → subramanyamdvss → Unsupnts

subramanyamdvss / Unsupnts

Unsupervised Neural Text Simplification

Programming Languages

python
139335 projects - #7 most used programming language

Labels

Projects that are alternatives of or similar to Unsupnts

Segan
Speech Enhancement Generative Adversarial Network in TensorFlow
Stars: ✭ 661 (+2773.91%)
Mutual labels:  gan
Anime Inpainting
An application tool of edge-connect, which can do anime inpainting and drawing. 动漫人物图片自动修复,去马赛克,填补,去瑕疵
Stars: ✭ 761 (+3208.7%)
Mutual labels:  gan
Lightning Bolts
Toolbox of models, callbacks, and datasets for AI/ML researchers.
Stars: ✭ 829 (+3504.35%)
Mutual labels:  gan
Passgan
A Deep Learning Approach for Password Guessing (https://arxiv.org/abs/1709.00440)
Stars: ✭ 704 (+2960.87%)
Mutual labels:  gan
Adversarialnetspapers
Awesome paper list with code about generative adversarial nets
Stars: ✭ 6,219 (+26939.13%)
Mutual labels:  gan
Pytorch Pretrained Biggan
🦋A PyTorch implementation of BigGAN with pretrained weights and conversion scripts.
Stars: ✭ 779 (+3286.96%)
Mutual labels:  gan
Ad examples
A collection of anomaly detection methods (iid/point-based, graph and time series) including active learning for anomaly detection/discovery, bayesian rule-mining, description for diversity/explanation/interpretability. Analysis of incorporating label feedback with ensemble and tree-based detectors. Includes adversarial attacks with Graph Convolutional Network.
Stars: ✭ 641 (+2686.96%)
Mutual labels:  gan
Csmri Refinement
Code for "Adversarial and Perceptual Refinement Compressed Sensing MRI Reconstruction"
Stars: ✭ 21 (-8.7%)
Mutual labels:  gan
Gans In Action
Companion repository to GANs in Action: Deep learning with Generative Adversarial Networks
Stars: ✭ 748 (+3152.17%)
Mutual labels:  gan
Tensorlayer
Deep Learning and Reinforcement Learning Library for Scientists and Engineers 🔥
Stars: ✭ 6,796 (+29447.83%)
Mutual labels:  gan
Machine Learning
머신러닝 입문자 혹은 스터디를 준비하시는 분들에게 도움이 되고자 만든 repository입니다. (This repository is intented for helping whom are interested in machine learning study)
Stars: ✭ 705 (+2965.22%)
Mutual labels:  gan
Context Encoder
[CVPR 2016] Unsupervised Feature Learning by Image Inpainting using GANs
Stars: ✭ 731 (+3078.26%)
Mutual labels:  gan
Musegan
An AI for Music Generation
Stars: ✭ 794 (+3352.17%)
Mutual labels:  gan
Adversarial video generation
A TensorFlow Implementation of "Deep Multi-Scale Video Prediction Beyond Mean Square Error" by Mathieu, Couprie & LeCun.
Stars: ✭ 662 (+2778.26%)
Mutual labels:  gan
Xdf Gan
A GAN for the generation of mock astronomical surveys
Stars: ✭ 17 (-26.09%)
Mutual labels:  gan
Pggan Pytorch
🔥🔥 PyTorch implementation of "Progressive growing of GANs (PGGAN)" 🔥🔥
Stars: ✭ 653 (+2739.13%)
Mutual labels:  gan
Instagan
InstaGAN: Instance-aware Image Translation (ICLR 2019)
Stars: ✭ 761 (+3208.7%)
Mutual labels:  gan
Advanced Deep Learning With Keras
Advanced Deep Learning with Keras, published by Packt
Stars: ✭ 917 (+3886.96%)
Mutual labels:  gan
Began Tensorflow
Tensorflow implementation of "BEGAN: Boundary Equilibrium Generative Adversarial Networks"
Stars: ✭ 904 (+3830.43%)
Mutual labels:  gan
Generative Models
Collection of generative models, e.g. GAN, VAE in Pytorch and Tensorflow.
Stars: ✭ 6,701 (+29034.78%)
Mutual labels:  gan

UnsupNTS: Unsupervised Neural Text Simplification

This is the original implementation of the Unsupervised Neural Text Simplification system and their semi-supervised variants mentioned in the ACL 2019 long paper:

Sai Surya, Abhijit Mishra, Anirban Laha, Parag Jain, and Karthik Sankaranarayanan. Unsupervised Neural Text Simplification arXiv preprint arXiv:1810.07931 (2018).

Training

Download tsdata.zip from link and extract

unzip tsdata.zip

tsdata.zip has

  • Unsupervised sets of easy and difficult set of sentences judged on readability ease scores.
  • Dict2vec embeddings trained on the above unsupervised sets.
  • 10k parallel pairs of difficult and simplified variants.
  • Test set and references - eight tab seperated references per each test sentence.

Train the models using

bash train.sh

train.sh has

  • UNTS system from unsupervised simplification data using the exact same settings described in the paper.
  • UNTS-10k system, using additional 10k supervised pairs of mixture of split-rephrase and simplification parallel pairs.
  • UNMT system on the unsupervised simplification data.
  • ablations on adversarial and separation/classifier losses.

For more details and additional options, run the above scripts with the --help flag. Alternatively, visit the ipynb in google colaboratory to reproduce the results. To access pretrained models visit link. The folder predictions has the generations from the pretrained models.

Note: Pretrained models were trained with pytorch 0.3.1 and may not exactly reproduce the result in pytorch 1.0.1 .

Generation and Evaluation of Simplifications

bash translate.sh

translate.sh is used for

  • Generating simplifications of test dataset.
  • Computing stand alone metrics such as Flesch readability ease score difference, Tree similarity and Document similarity metrics.
  • Computing SARI, BLEU and Word-diff metrics.

Acknowledgements

A large portion of this repo is borrowed from the following repos: https://github.com/artetxem/undreamt and https://github.com/senisioi/NeuralTextSimplification.

If you use our system for academic research, please cite the following paper:

@inproceedings{surya-etal-2019-unsupervised,
    title = "Unsupervised Neural Text Simplification",
    author = "Surya, Sai  and
      Mishra, Abhijit  and
      Laha, Anirban  and
      Jain, Parag  and
      Sankaranarayanan, Karthik",
    booktitle = "Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics",
    month = jul,
    year = "2019",
    address = "Florence, Italy",
    publisher = "Association for Computational Linguistics",
    url = "https://www.aclweb.org/anthology/P19-1198",
    doi = "10.18653/v1/P19-1198",
    pages = "2058--2068"
}
Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].