All Projects → senisioi → NeuralTextSimplification

senisioi / NeuralTextSimplification

Licence: other
Exploring Neural Text Simplification

Programming Languages

lua
6591 projects
python
139335 projects - #7 most used programming language
shell
77523 projects

Projects that are alternatives of or similar to NeuralTextSimplification

SRB
Code for "Improving Semantic Relevance for Sequence-to-Sequence Learning of Chinese Social Media Text Summarization"
Stars: ✭ 41 (-35.94%)
Mutual labels:  simplification, seq2seq
classy
classy is a simple-to-use library for building high-performance Machine Learning models in NLP.
Stars: ✭ 61 (-4.69%)
Mutual labels:  seq2seq
cnn-seq2seq
No description or website provided.
Stars: ✭ 39 (-39.06%)
Mutual labels:  seq2seq
GAN-RNN Timeseries-imputation
Recurrent GAN for imputation of time series data. Implemented in TensorFlow 2 on Wikipedia Web Traffic Forecast dataset from Kaggle.
Stars: ✭ 107 (+67.19%)
Mutual labels:  seq2seq
Embedding
Embedding模型代码和学习笔记总结
Stars: ✭ 25 (-60.94%)
Mutual labels:  seq2seq
ttslearn
ttslearn: Library for Pythonで学ぶ音声合成 (Text-to-speech with Python)
Stars: ✭ 158 (+146.88%)
Mutual labels:  seq2seq
RNNSearch
An implementation of attention-based neural machine translation using Pytorch
Stars: ✭ 43 (-32.81%)
Mutual labels:  seq2seq
ai-visual-storytelling-seq2seq
Implementation of seq2seq model for Visual Storytelling Challenge (VIST) http://visionandlanguage.net/VIST/index.html
Stars: ✭ 50 (-21.87%)
Mutual labels:  seq2seq
DLCV2018SPRING
Deep Learning for Computer Vision (CommE 5052) in NTU
Stars: ✭ 38 (-40.62%)
Mutual labels:  seq2seq
deep-molecular-optimization
Molecular optimization by capturing chemist’s intuition using the Seq2Seq with attention and the Transformer
Stars: ✭ 60 (-6.25%)
Mutual labels:  seq2seq
pytorch-transformer-chatbot
PyTorch v1.2에서 생긴 Transformer API 를 이용한 간단한 Chitchat 챗봇
Stars: ✭ 44 (-31.25%)
Mutual labels:  seq2seq
Base-On-Relation-Method-Extract-News-DA-RNN-Model-For-Stock-Prediction--Pytorch
基於關聯式新聞提取方法之雙階段注意力機制模型用於股票預測
Stars: ✭ 33 (-48.44%)
Mutual labels:  seq2seq
MoChA-pytorch
PyTorch Implementation of "Monotonic Chunkwise Attention" (ICLR 2018)
Stars: ✭ 65 (+1.56%)
Mutual labels:  seq2seq
Video-Cap
🎬 Video Captioning: ICCV '15 paper implementation
Stars: ✭ 44 (-31.25%)
Mutual labels:  seq2seq
neural-chat
An AI chatbot using seq2seq
Stars: ✭ 30 (-53.12%)
Mutual labels:  seq2seq
AI physicist
AI Physicist, a paradigm with algorithms for learning theories from data, by Wu and Tegmark (2019)
Stars: ✭ 23 (-64.06%)
Mutual labels:  simplification
transformer
Neutron: A pytorch based implementation of Transformer and its variants.
Stars: ✭ 60 (-6.25%)
Mutual labels:  seq2seq
karnaugh-map-simplifier
Karnaugh map simplification software, used to visually simplify boolean expressions
Stars: ✭ 33 (-48.44%)
Mutual labels:  simplification
torch-asg
Auto Segmentation Criterion (ASG) implemented in pytorch
Stars: ✭ 42 (-34.37%)
Mutual labels:  seq2seq
chatbot
🤖️ 基于 PyTorch 的任务型聊天机器人(支持私有部署和 docker 部署的 Chatbot)
Stars: ✭ 77 (+20.31%)
Mutual labels:  seq2seq

Exploring Neural Text Simplification

Abstract

We present the first attempt at using sequence to sequence neural networks to model text simplification (TS). Unlike the previously proposed automated methods, our neural text simplification (NTS) systems are able to simultaneously perform lexical simplification and content reduction. An extensive human evaluation of the output has shown that NTS systems achieve good grammaticality and meaning preservation of output sentences and higher level of simplification than the state-of-the-art automated TS systems. We train our models on the Wikipedia corpus containing good and good partial alignments.

	@InProceedings{neural-text-simplification,
	  author    = {Sergiu Nisioi and Sanja Štajner and Simone Paolo Ponzetto and Liviu P. Dinu},
	  title     = {Exploring Neural Text Simplification Models},
	  booktitle = {{ACL} {(2)}},
	  publisher = {The Association for Computational Linguistics},
	  year      = {2017}
	}

Simplify Text | Generate Predictions (no GPUs needed)

  1. OpenNMT dependencies
    1. Install Torch
    2. Install additional packages:
    luarocks install tds
  2. Checkout this repository including the submodules:
   git clone --recursive https://github.com/senisioi/NeuralTextSimplification.git
  1. Download the pre-trained released models NTS and NTS-w2v (NOTE: when using the released pre-trained models, due to recent changes in third party software, the output of our systems might not be identical to the one reported in the paper.)
   python src/download_models.py ./models
  1. Run translate.sh from the scripts dir:
   cd src/scripts
   ./translate.sh
  1. Check the predictions in the results directory:
   cd ../../results_NTS
  1. Run automatic evaluation metrics
    1. Install the python requirements (only nltk is needed)
       pip install -r src/requirements.txt
    1. Run the evaluate script
       python src/evaluate.py ./data/test.en ./data/references/references.tsv ./predictions/

The Content of this Repository

./src

  • download_models.py a script to download the pre-trained models. The models are released to be usable on machines with or without GPUs. They can't be used to continue the training session. In case the download script fails, you may use the direct links for NTS and NTS-w2v
  • train_word2vec.py a script that creates a word2vec model from a local corpus, using gensim
  • SARI.py a copy of the SARI implementation
  • evaluate.py evaluates BLEU and SARI scores given a source file, a directory of predictions and a reference file in tsv format
  • ./scripts - contains some of our scripts that we used to preprocess the data, output translations, and create the concatenated embeddings
  • ./patch - the patch with some changes that need to be applied, in case you may want to use the latest checkout of OpenNMT. Alternatively, you may use our forked code which comes directly as a submodule.

./configs

Contains the OpenNMT config file. To train, please update the config file with the appropriate data on your local system and run

	th train -config $PATH_TO_THIS_DIR/configs/NTS.cfg

./predictions

Contains predictions from previous systems (Wubben et al., 2012), (Glavas and Stajner, 2015), and (Xu et al., 2016), and the generated predictions of the NTS models reported in the paper:

  • NTS_default_b5_h1 - the default model, beam size 5, hypothesis 1

  • NTS_BLEU_b12_h1 - the BLEU best ranked model, beam size 12, hypothesis 1

  • NTS_SARI_b5_h2 - the SARI best ranked model, beam size 12, hypothesis 1

  • NTS-w2v_default_b5_h1 - the default model, beam size 5, hypothesis 1

  • NTS-w2v_BLEU_b12_h1 - the BLEU best ranked model, beam size 12, hypothesis 1

  • NTS-w2v_SARI_b12_h2 - the SARI best ranked model, beam size 12, hypothesis 2

./data

Contains the training, testing, and reference sentences used to train and evaluate our models.

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].