All Projects → mandarjoshi90 → pair2vec

mandarjoshi90 / pair2vec

Licence: Apache-2.0 license
pair2vec: Compositional Word-Pair Embeddings for Cross-Sentence Inference

Programming Languages

python
139335 projects - #7 most used programming language
shell
77523 projects

Projects that are alternatives of or similar to pair2vec

Word2vec Win32
A word2vec port for Windows.
Stars: ✭ 41 (-33.87%)
Mutual labels:  word-embeddings, representation-learning
Knowledge Graphs
A collection of research on knowledge graphs
Stars: ✭ 845 (+1262.9%)
Mutual labels:  question-answering, representation-learning
backprop
Backprop makes it simple to use, finetune, and deploy state-of-the-art ML models.
Stars: ✭ 229 (+269.35%)
Mutual labels:  question-answering
word-benchmarks
Benchmarks for intrinsic word embeddings evaluation.
Stars: ✭ 45 (-27.42%)
Mutual labels:  word-embeddings
GrailQA
No description or website provided.
Stars: ✭ 72 (+16.13%)
Mutual labels:  question-answering
REGAL
Representation learning-based graph alignment based on implicit matrix factorization and structural embeddings
Stars: ✭ 78 (+25.81%)
Mutual labels:  representation-learning
GLOM-TensorFlow
An attempt at the implementation of GLOM, Geoffrey Hinton's paper for emergent part-whole hierarchies from data
Stars: ✭ 32 (-48.39%)
Mutual labels:  representation-learning
S-WMD
Code for Supervised Word Mover's Distance (SWMD)
Stars: ✭ 90 (+45.16%)
Mutual labels:  word-embeddings
Arabic-Word-Embeddings-Word2vec
Arabic Word Embeddings Word2vec
Stars: ✭ 26 (-58.06%)
Mutual labels:  word-embeddings
M-NMF
An implementation of "Community Preserving Network Embedding" (AAAI 2017)
Stars: ✭ 119 (+91.94%)
Mutual labels:  representation-learning
EBIM-NLI
Enhanced BiLSTM Inference Model for Natural Language Inference
Stars: ✭ 24 (-61.29%)
Mutual labels:  natural-language-inference
contextualLSTM
Contextual LSTM for NLP tasks like word prediction and word embedding creation for Deep Learning
Stars: ✭ 28 (-54.84%)
Mutual labels:  word-embeddings
banglabert
This repository contains the official release of the model "BanglaBERT" and associated downstream finetuning code and datasets introduced in the paper titled "BanglaBERT: Language Model Pretraining and Benchmarks for Low-Resource Language Understanding Evaluation in Bangla" accpeted in Findings of the Annual Conference of the North American Chap…
Stars: ✭ 186 (+200%)
Mutual labels:  natural-language-inference
awesome-graph-self-supervised-learning
Awesome Graph Self-Supervised Learning
Stars: ✭ 805 (+1198.39%)
Mutual labels:  representation-learning
RECCON
This repository contains the dataset and the PyTorch implementations of the models from the paper Recognizing Emotion Cause in Conversations.
Stars: ✭ 126 (+103.23%)
Mutual labels:  natural-language-inference
TCE
This repository contains the code implementation used in the paper Temporally Coherent Embeddings for Self-Supervised Video Representation Learning (TCE).
Stars: ✭ 51 (-17.74%)
Mutual labels:  representation-learning
MTL-AQA
What and How Well You Performed? A Multitask Learning Approach to Action Quality Assessment [CVPR 2019]
Stars: ✭ 38 (-38.71%)
Mutual labels:  representation-learning
dasem
Danish Semantic analysis
Stars: ✭ 17 (-72.58%)
Mutual labels:  word-embeddings
reprieve
A library for evaluating representations.
Stars: ✭ 68 (+9.68%)
Mutual labels:  representation-learning
amr
Official adversarial mixup resynthesis repository
Stars: ✭ 31 (-50%)
Mutual labels:  representation-learning

pair2vec: Compositional Word-Pair Embeddings for Cross-Sentence Inference

Introduction

This repository contains the code for replicating results from

Getting Started

  • Install python3 requirements: pip install -r requirements.txt

Using pretrained pair2vec embeddings

  • Download pretrained pair2vec: ./download_pair2vec.sh
    • If you want to reproduce results from the paper on QA/NLI, please use the following:
      • Download and extract the pretrained models tar file
      • Run evaluation:
    python -m allennlp.run evaluate [--output-file OUTPUT_FILE]
                                 --cuda-device 0
                                 --include-package endtasks
                                 ARCHIVE_FILE INPUT_FILE
    
    • If you want to train your own QA/NLI model:
    python -m allennlp.run train <config_file> -s <serialization_dir> --include-package endtasks
    

See the experiments directory for relevant config files.

Training your own embeddings

  • Download the preprocessed corpus if you want to train pair2vec from scratch: ./download_corpus.sh
  • Training: This starts the training process which typically takes 7-10 days. It takes in a config file and a directory to save checkpoints.
python -m embeddings.train --config experiments/pair2vec_train.json --save_path <directory>

Miscellaneous

  • If you use the code, please cite the following paper
@inproceedings{joshi-etal-2019-pair2vec,
    title = "pair2vec: Compositional Word-Pair Embeddings for Cross-Sentence Inference",
    author = "Joshi, Mandar  and
      Choi, Eunsol  and
      Levy, Omer  and
      Weld, Daniel  and
      Zettlemoyer, Luke",
    booktitle = "Proceedings of the 2019 Conference of the North {A}merican Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers)",
    month = jun,
    year = "2019",
    address = "Minneapolis, Minnesota",
    publisher = "Association for Computational Linguistics",
    url = "https://www.aclweb.org/anthology/N19-1362",
    pages = "3597--3608"
}
Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].