All Projects → KelleyYin → Cross-lingual-Summarization

KelleyYin / Cross-lingual-Summarization

Licence: other
Zero-Shot Cross-Lingual Abstractive Sentence Summarization through Teaching Generation and Attention

Programming Languages

python
139335 projects - #7 most used programming language
xBase
55 projects
Charity
2 projects
shell
77523 projects

Projects that are alternatives of or similar to Cross-lingual-Summarization

BMT
Source code for "Bi-modal Transformer for Dense Video Captioning" (BMVC 2020)
Stars: ✭ 192 (+585.71%)
Mutual labels:  transformer
bytekit
Java 字节操作的工具库(不是字节码的工具库)
Stars: ✭ 40 (+42.86%)
Mutual labels:  transformer
classmanager-student-teacher-portal
A Student-Teacher Portal built using HTML, CSS, Python and Django
Stars: ✭ 155 (+453.57%)
Mutual labels:  teacher-student-learning
densecap
Dense video captioning in PyTorch
Stars: ✭ 37 (+32.14%)
Mutual labels:  transformer
MASTER-pytorch
Code for the paper "MASTER: Multi-Aspect Non-local Network for Scene Text Recognition" (Pattern Recognition 2021)
Stars: ✭ 263 (+839.29%)
Mutual labels:  transformer
keras-vision-transformer
The Tensorflow, Keras implementation of Swin-Transformer and Swin-UNET
Stars: ✭ 91 (+225%)
Mutual labels:  transformer
Transformers-RL
An easy PyTorch implementation of "Stabilizing Transformers for Reinforcement Learning"
Stars: ✭ 107 (+282.14%)
Mutual labels:  transformer
php-serializer
Serialize PHP variables, including objects, in any format. Support to unserialize it too.
Stars: ✭ 47 (+67.86%)
Mutual labels:  transformer
t5-japanese
Codes to pre-train Japanese T5 models
Stars: ✭ 39 (+39.29%)
Mutual labels:  transformer
ViTs-vs-CNNs
[NeurIPS 2021]: Are Transformers More Robust Than CNNs? (Pytorch implementation & checkpoints)
Stars: ✭ 145 (+417.86%)
Mutual labels:  transformer
vietnamese-roberta
A Robustly Optimized BERT Pretraining Approach for Vietnamese
Stars: ✭ 22 (-21.43%)
Mutual labels:  transformer
TianChi AIEarth
TianChi AIEarth Contest Solution
Stars: ✭ 57 (+103.57%)
Mutual labels:  transformer
Representation-Learning-for-Information-Extraction
Pytorch implementation of Paper by Google Research - Representation Learning for Information Extraction from Form-like Documents.
Stars: ✭ 82 (+192.86%)
Mutual labels:  transformer
SegSwap
(CVPRW 2022) Learning Co-segmentation by Segment Swapping for Retrieval and Discovery
Stars: ✭ 46 (+64.29%)
Mutual labels:  transformer
fastT5
⚡ boost inference speed of T5 models by 5x & reduce the model size by 3x.
Stars: ✭ 421 (+1403.57%)
Mutual labels:  transformer
sb-nmt
Code for Synchronous Bidirectional Neural Machine Translation (SB-NMT)
Stars: ✭ 66 (+135.71%)
Mutual labels:  transformer
Awesome-low-level-vision-resources
A curated list of resources for Low-level Vision Tasks
Stars: ✭ 35 (+25%)
Mutual labels:  transformer
project-code-py
Leetcode using AI
Stars: ✭ 100 (+257.14%)
Mutual labels:  transformer
ru-dalle
Generate images from texts. In Russian
Stars: ✭ 1,606 (+5635.71%)
Mutual labels:  transformer
R-MeN
Transformer-based Memory Networks for Knowledge Graph Embeddings (ACL 2020) (Pytorch and Tensorflow)
Stars: ✭ 74 (+164.29%)
Mutual labels:  transformer

Cross-Lingual Abstractive Sentence Summarization(CL-ASSUM)

Introduction

We implemented CL-ASSUM on fairseq. In this repo, it contains of four parts.

  • Transformer
  • Teaching-Generation
  • Teaching-Attention
  • Teaching-Generation-Attention

Teacher models

Before staring the experiment, you should first use Transformer to train the teacher model of NMT model and momolingual summarization model.

Please refer to Transformer for more deatils.

Teaching-Generation

Teaching-Attention

Teaching-Generation-Attention

Evaluation sets of CL-ASSUM

The test-data file contains evaluation sets of CL-ASSUM, which is built by manual translation.

Requirements and Installation

  • A PyTorch installation
  • For training new models, you'll also need an NVIDIA GPU and NCCL
  • Python version 3.6
  • PyTorch version >= 0.4.0.

Cross-Lingual Test Set

In our experiments, we manually translate the English sentences into the Chinese sentences for the validation and evaluation sets of Gigaword and DUC2004.

License

Reference

If you find CL-ASSUM useful in your work, you can cite this paper as below:

@inproceedings{duan-etal-2019-zero,
    title = "Zero-Shot Cross-Lingual Abstractive Sentence Summarization through Teaching Generation and Attention",
    author = "Duan, Xiangyu  and Yin, Mingming  and Zhang, Min  and Chen, Boxing  and Luo, Weihua",
    booktitle = "Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics",
    month = jul,
    year = "2019",
    address = "Florence, Italy",
    publisher = "Association for Computational Linguistics",
    url = "https://www.aclweb.org/anthology/P19-1305",
    doi = "10.18653/v1/P19-1305",
    pages = "3162--3172",
   }
Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].