All Projects → Hironsan → Awesome Embedding Models

Hironsan / Awesome Embedding Models

Licence: mit
A curated list of awesome embedding models tutorials, projects and communities.

Programming Languages

Jupyter Notebook
11667 projects

Projects that are alternatives of or similar to Awesome Embedding Models

Deep Math Machine Learning.ai
A blog which talks about machine learning, deep learning algorithms and the Math. and Machine learning algorithms written from scratch.
Stars: ✭ 173 (-88.36%)
Mutual labels:  jupyter-notebook, natural-language-processing, word2vec
Nlp In Practice
Starter code to solve real world text data problems. Includes: Gensim Word2Vec, phrase embeddings, Text Classification with Logistic Regression, word count with pyspark, simple text preprocessing, pre-trained embeddings and more.
Stars: ✭ 790 (-46.84%)
Mutual labels:  jupyter-notebook, natural-language-processing, word2vec
Germanwordembeddings
Toolkit to obtain and preprocess german corpora, train models using word2vec (gensim) and evaluate them with generated testsets
Stars: ✭ 189 (-87.28%)
Mutual labels:  jupyter-notebook, natural-language-processing, word2vec
Magnitude
A fast, efficient universal vector embedding utility package.
Stars: ✭ 1,394 (-6.19%)
Mutual labels:  natural-language-processing, word2vec, embeddings
Practical 1
Oxford Deep NLP 2017 course - Practical 1: word2vec
Stars: ✭ 220 (-85.2%)
Mutual labels:  jupyter-notebook, natural-language-processing, word2vec
Deeplearning Nlp Models
A small, interpretable codebase containing the re-implementation of a few "deep" NLP models in PyTorch. Colab notebooks to run with GPUs. Models: word2vec, CNNs, transformer, gpt.
Stars: ✭ 64 (-95.69%)
Mutual labels:  jupyter-notebook, word2vec, embeddings
Math And Ml Notes
Books, papers and links to latest research in ML/AI
Stars: ✭ 76 (-94.89%)
Mutual labels:  jupyter-notebook, papers
Ja.text8
Japanese text8 corpus for word embedding.
Stars: ✭ 79 (-94.68%)
Mutual labels:  natural-language-processing, word2vec
Glove As A Tensorflow Embedding Layer
Taking a pretrained GloVe model, and using it as a TensorFlow embedding weight layer **inside the GPU**. Therefore, you only need to send the index of the words through the GPU data transfer bus, reducing data transfer overhead.
Stars: ✭ 85 (-94.28%)
Mutual labels:  jupyter-notebook, word2vec
Dict2vec
Dict2vec is a framework to learn word embeddings using lexical dictionaries.
Stars: ✭ 91 (-93.88%)
Mutual labels:  word2vec, embeddings
Python nlp tutorial
This repository provides everything to get started with Python for Text Mining / Natural Language Processing (NLP)
Stars: ✭ 72 (-95.15%)
Mutual labels:  jupyter-notebook, natural-language-processing
Spark Nlp Models
Models and Pipelines for the Spark NLP library
Stars: ✭ 88 (-94.08%)
Mutual labels:  jupyter-notebook, natural-language-processing
Experiments
Some research experiments
Stars: ✭ 95 (-93.61%)
Mutual labels:  jupyter-notebook, word2vec
Nlp Tutorial
Natural Language Processing Tutorial for Deep Learning Researchers
Stars: ✭ 9,895 (+565.88%)
Mutual labels:  jupyter-notebook, natural-language-processing
Course Computational Literary Analysis
Course materials for Introduction to Computational Literary Analysis, taught at UC Berkeley in Summer 2018, 2019, and 2020, and at Columbia University in Fall 2020.
Stars: ✭ 74 (-95.02%)
Mutual labels:  jupyter-notebook, natural-language-processing
Turkish Bert Nlp Pipeline
Bert-base NLP pipeline for Turkish, Ner, Sentiment Analysis, Question Answering etc.
Stars: ✭ 85 (-94.28%)
Mutual labels:  jupyter-notebook, natural-language-processing
Nlp Tutorial
A list of NLP(Natural Language Processing) tutorials
Stars: ✭ 1,188 (-20.05%)
Mutual labels:  jupyter-notebook, natural-language-processing
Textclustering
Stars: ✭ 89 (-94.01%)
Mutual labels:  jupyter-notebook, word2vec
Codesearchnet
Datasets, tools, and benchmarks for representation learning of code.
Stars: ✭ 1,378 (-7.27%)
Mutual labels:  jupyter-notebook, natural-language-processing
Awesome Pytorch List Cnversion
Awesome-pytorch-list 翻译工作进行中......
Stars: ✭ 1,361 (-8.41%)
Mutual labels:  jupyter-notebook, papers

awesome-embedding-modelsAwesome

A curated list of awesome embedding models tutorials, projects and communities. Please feel free to pull requests to add links.

Table of Contents

Papers

Word Embeddings

Word2vec, GloVe, FastText

  • Efficient Estimation of Word Representations in Vector Space (2013), T. Mikolov et al. [pdf]
  • Distributed Representations of Words and Phrases and their Compositionality (2013), T. Mikolov et al. [pdf]
  • word2vec Parameter Learning Explained (2014), Xin Rong [pdf]
  • word2vec Explained: deriving Mikolov et al.'s negative-sampling word-embedding method (2014), Yoav Goldberg, Omer Levy [pdf]
  • GloVe: Global Vectors for Word Representation (2014), J. Pennington et al. [pdf]
  • Improving Word Representations via Global Context and Multiple Word Prototypes (2012), EH Huang et al. [pdf]
  • Enriching Word Vectors with Subword Information (2016), P. Bojanowski et al. [pdf]
  • Bag of Tricks for Efficient Text Classification (2016), A. Joulin et al. [pdf]

Language Model

  • Semi-supervised sequence tagging with bidirectional language models (2017), Peters, Matthew E., et al. [pdf]
  • Deep contextualized word representations (2018), Peters, Matthew E., et al. [pdf]
  • Contextual String Embeddings for Sequence Labeling (2018), Akbik, Alan, Duncan Blythe, and Roland Vollgraf. [pdf]
  • BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding (2018), [pdf]

Embedding Enhancement

  • Sentence Embedding:Learning Semantic Sentence Embeddings using Pair-wise Discriminator(2018),Patro et al.[Project Page] [Paper]
  • Retrofitting Word Vectors to Semantic Lexicons (2014), M. Faruqui et al. [pdf]
  • Better Word Representations with Recursive Neural Networks for Morphology (2013), T.Luong et al. [pdf]
  • Dependency-Based Word Embeddings (2014), Omer Levy, Yoav Goldberg [pdf]
  • Not All Neural Embeddings are Born Equal (2014), F. Hill et al. [pdf]
  • Two/Too Simple Adaptations of Word2Vec for Syntax Problems (2015), W. Ling[pdf]

Comparing count-based vs predict-based method

  • Linguistic Regularities in Sparse and Explicit Word Representations (2014), Omer Levy, Yoav Goldberg[pdf]
  • Don’t count, predict! A systematic comparison of context-counting vs. context-predicting semantic vectors (2014), M. Baroni [pdf]
  • Improving Distributional Similarity with Lessons Learned from Word Embeddings (2015), Omer Levy [pdf]

Evaluation, Analysis

  • Evaluation methods for unsupervised word embeddings (2015), T. Schnabel [pdf]
  • Intrinsic Evaluation of Word Vectors Fails to Predict Extrinsic Performance (2016), B. Chiu [pdf]
  • Problems With Evaluation of Word Embeddings Using Word Similarity Tasks (2016), M. Faruqui [pdf]
  • Improving Reliability of Word Similarity Evaluation by Redesigning Annotation Task and Performance Measure (2016), Oded Avraham, Yoav Goldberg [pdf]
  • Evaluating Word Embeddings Using a Representative Suite of Practical Tasks (2016), N. Nayak [pdf]

Phrase, Sentence and Document Embeddings

Sentence

Document

Sense Embeddings

Neural Language Models

Researchers

Courses and Lectures

Datasets

Training

Evaluation

Pre-Trained Language Models

Below is pre-trained ELMo models. Adding ELMo to existing NLP systems significantly improves the state-of-the-art for every considered task.

Below is pre-trained sent2vec models.

Pre-Trained Word Vectors

Convenient downloader for pre-trained word vectors:

Links for pre-trained word vectors:

Implementations and Tools

Word2vec

GloVe

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].