All Projects → EagleW → Paperrobot

EagleW / Paperrobot

Licence: mit
Code for PaperRobot: Incremental Draft Generation of Scientific Ideas

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to Paperrobot

Describing a knowledge base
Code for Describing a Knowledge Base
Stars: ✭ 42 (-88.71%)
Mutual labels:  datasets, attention-mechanism, text-generation, natural-language-generation, generation
Writing-editing-Network
Code for Paper Abstract Writing through Editing Mechanism
Stars: ✭ 72 (-80.65%)
Mutual labels:  text-generation, generation, datasets, attention-mechanism
factedit
🧐 Code & Data for Fact-based Text Editing (Iso et al; ACL 2020)
Stars: ✭ 16 (-95.7%)
Mutual labels:  text-generation, natural-language-generation
Attention Mechanisms
Implementations for a family of attention mechanisms, suitable for all kinds of natural language processing tasks and compatible with TensorFlow 2.0 and Keras.
Stars: ✭ 203 (-45.43%)
Mutual labels:  attention-mechanism, text-generation
3d Pointcloud
Papers and Datasets about Point Cloud.
Stars: ✭ 179 (-51.88%)
Mutual labels:  datasets, generation
PlanSum
[AAAI2021] Unsupervised Opinion Summarization with Content Planning
Stars: ✭ 25 (-93.28%)
Mutual labels:  text-generation, natural-language-generation
Entity2Topic
[NAACL2018] Entity Commonsense Representation for Neural Abstractive Summarization
Stars: ✭ 20 (-94.62%)
Mutual labels:  text-generation, natural-language-generation
keras-deep-learning
Various implementations and projects on CNN, RNN, LSTM, GAN, etc
Stars: ✭ 22 (-94.09%)
Mutual labels:  text-generation, attention-mechanism
Kenlg Reading
Reading list for knowledge-enhanced text generation, with a survey
Stars: ✭ 257 (-30.91%)
Mutual labels:  text-generation, natural-language-generation
Accelerated Text
Accelerated Text is a no-code natural language generation platform. It will help you construct document plans which define how your data is converted to textual descriptions varying in wording and structure.
Stars: ✭ 256 (-31.18%)
Mutual labels:  text-generation, natural-language-generation
Awesome Segmentation Saliency Dataset
A collection of some datasets for segmentation / saliency detection. Welcome to PR...😄
Stars: ✭ 315 (-15.32%)
Mutual labels:  datasets
Yolo Multi Backbones Attention
Model Compression—YOLOv3 with multi lightweight backbones(ShuffleNetV2 HuaWei GhostNet), attention, prune and quantization
Stars: ✭ 317 (-14.78%)
Mutual labels:  attention-mechanism
Nlp Conference Compendium
Compendium of the resources available from top NLP conferences.
Stars: ✭ 349 (-6.18%)
Mutual labels:  natural-language-generation
Question generation
Neural question generation using transformers
Stars: ✭ 356 (-4.3%)
Mutual labels:  natural-language-generation
Seq2seq chatbot
基于seq2seq模型的简单对话系统的tf实现,具有embedding、attention、beam_search等功能,数据集是Cornell Movie Dialogs
Stars: ✭ 308 (-17.2%)
Mutual labels:  attention-mechanism
Action Recognition Visual Attention
Action recognition using soft attention based deep recurrent neural networks
Stars: ✭ 350 (-5.91%)
Mutual labels:  attention-mechanism
Tensorflow end2end speech recognition
End-to-End speech recognition implementation base on TensorFlow (CTC, Attention, and MTL training)
Stars: ✭ 305 (-18.01%)
Mutual labels:  attention-mechanism
Medical Datasets
tracking medical datasets, with a focus on medical imaging
Stars: ✭ 296 (-20.43%)
Mutual labels:  datasets
Cleora
Cleora AI is a general-purpose model for efficient, scalable learning of stable and inductive entity embeddings for heterogeneous relational data.
Stars: ✭ 303 (-18.55%)
Mutual labels:  datasets
Animal Matting
Github repository for the paper End-to-end Animal Image Matting
Stars: ✭ 363 (-2.42%)
Mutual labels:  datasets

PaperRobot: Incremental Draft Generation of Scientific Ideas

PaperRobot: Incremental Draft Generation of Scientific Ideas [Sample Output]

Accepted by 57th Annual Meeting of the Association for Computational Linguistics (ACL 2019)

Table of Contents

Overview

Photo

Requirements

Environment:

Pacakges

You can click the following links for detailed installation instructions.

Data:

  • PubMed Paper Reading Dataset This dataset gathers 14,857 entities, 133 relations, and entities corresponding tokenized text from PubMed. It contains 875,698 training pairs, 109,462 development pairs, and 109,462 test pairs.

  • PubMed Term, Abstract, Conclusion, Title Dataset This dataset gathers three types of pairs: Title-to-Abstract (Training: 22,811/Development: 2095/Test: 2095), Abstract-to-Conclusion and Future work (Training: 22,811/Development: 2095/Test: 2095), Conclusion and Future work-to-Title (Training: 15,902/Development: 2095/Test: 2095) from PubMed. Each pair contains a pair of input and output as well as the corresponding terms(from original KB and link prediction results).

Quickstart

Existing paper reading

CAUTION!! Because the dataset is quite large, the training and evaluation of link prediction model will be pretty slow.

Preprocessing:

Download and unzip the paper_reading.zip from PubMed Paper Reading Dataset . Put paper_reading folder under the Existing paper reading folder.

Training

Hyperparameter can be adjusted as follows: For example, if you want to change the number of hidden unit to 6, you can append --hidden 6 after train.py

python train.py

To resume training, you can apply the following command and put the previous model path after the --model

python train.py --cont --model models/GATA/best_dev_model.pth.tar

Test

Put the finished model path after the --model The test.py will provide the ranking score for the test set.

python test.py --model models/GATA/best_dev_model.pth.tar

New paper writing

Preprocessing:

Download and unzip the data_pubmed_writing.zip from PubMed Term, Abstract, Conclusion, Title Dataset . Put data folder under the New paper writing folder.

Training

Put the type of data after the --data_path. For example, if you want to train an abstract model, put data/pubmed_abstract after --data_path. Put the model directory after the --model_dp

python train.py --data_path data/pubmed_abstract --model_dp abstract_model/

To resume training, you can apply the following command and put the previous model path after the --model

python train.py --data_path data/pubmed_abstract --cont --model abstract_model/memory/best_dev_model.pth.tar

For more other options, please check the code.

Test

Put the finished model path after the --model The test.py will provide the score for the test set.

python test.py --data_path data/pubmed_abstract --model abstract_model/memory/best_dev_model.pth.tar

Predict an instance

Put the finished model path after the --model The input.py will provide the prediction for customized input.

python input.py --data_path data/pubmed_abstract --model abstract_model/memory/best_dev_model.pth.tar

Citation

@inproceedings{wang-etal-2019-paperrobot,
    title = "{P}aper{R}obot: Incremental Draft Generation of Scientific Ideas",
    author = "Wang, Qingyun  and
      Huang, Lifu  and
      Jiang, Zhiying  and
      Knight, Kevin  and
      Ji, Heng  and
      Bansal, Mohit  and
      Luan, Yi",
    booktitle = "Proceedings of the 57th Conference of the Association for Computational Linguistics",
    month = jul,
    year = "2019",
    address = "Florence, Italy",
    publisher = "Association for Computational Linguistics",
    url = "https://www.aclweb.org/anthology/P19-1191",
    pages = "1980--1991"
}
Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].