KoanA word2vec negative sampling implementation with correct CBOW update.
Stars: ✭ 232 (+1060%)
Datastories Semeval2017 Task4Deep-learning model presented in "DataStories at SemEval-2017 Task 4: Deep LSTM with Attention for Message-level and Topic-based Sentiment Analysis".
Stars: ✭ 184 (+820%)
word-benchmarksBenchmarks for intrinsic word embeddings evaluation.
Stars: ✭ 45 (+125%)
HiCECode for ACL'19 "Few-Shot Representation Learning for Out-Of-Vocabulary Words"
Stars: ✭ 56 (+180%)
Elmo TutorialA short tutorial on Elmo training (Pre trained, Training on new data, Incremental training)
Stars: ✭ 145 (+625%)
compress-fasttextTools for shrinking fastText models (in gensim format)
Stars: ✭ 124 (+520%)
ShallowlearnAn experiment about re-implementing supervised learning models based on shallow neural network approaches (e.g. fastText) with some additional exclusive features and nice API. Written in Python and fully compatible with Scikit-learn.
Stars: ✭ 196 (+880%)
QuestionClusteringClasificador de preguntas escrito en python 3 que fue implementado en el siguiente vídeo: https://youtu.be/qnlW1m6lPoY
Stars: ✭ 15 (-25%)
LftmImproving topic models LDA and DMM (one-topic-per-document model for short texts) with word embeddings (TACL 2015)
Stars: ✭ 168 (+740%)
S-WMDCode for Supervised Word Mover's Distance (SWMD)
Stars: ✭ 90 (+350%)
two-stream-cnnA two-stream convolutional neural network for learning abitrary similarity functions over two sets of training data
Stars: ✭ 24 (+20%)
ScattertextBeautiful visualizations of how language differs among document types.
Stars: ✭ 1,722 (+8510%)
JoSH[KDD 2020] Hierarchical Topic Mining via Joint Spherical Tree and Text Embedding
Stars: ✭ 55 (+175%)
Pytorch Sentiment AnalysisTutorials on getting started with PyTorch and TorchText for sentiment analysis.
Stars: ✭ 3,209 (+15945%)
lda2vecMixing Dirichlet Topic Models and Word Embeddings to Make lda2vec from this paper https://arxiv.org/abs/1605.02019
Stars: ✭ 27 (+35%)
Question GenerationGenerating multiple choice questions from text using Machine Learning.
Stars: ✭ 227 (+1035%)
pair2vecpair2vec: Compositional Word-Pair Embeddings for Cross-Sentence Inference
Stars: ✭ 62 (+210%)
GermanwordembeddingsToolkit to obtain and preprocess german corpora, train models using word2vec (gensim) and evaluate them with generated testsets
Stars: ✭ 189 (+845%)
NTUA-slp-nlp💻Speech and Natural Language Processing (SLP & NLP) Lab Assignments for ECE NTUA
Stars: ✭ 19 (-5%)
DebiasweRemove problematic gender bias from word embeddings.
Stars: ✭ 175 (+775%)
dasemDanish Semantic analysis
Stars: ✭ 17 (-15%)
MimickCode for Mimicking Word Embeddings using Subword RNNs (EMNLP 2017)
Stars: ✭ 152 (+660%)
SiameseCBOWImplementation of Siamese CBOW using keras whose backend is tensorflow.
Stars: ✭ 14 (-30%)
fuzzymaxCode for the paper: Don't Settle for Average, Go for the Max: Fuzzy Sets and Max-Pooled Word Vectors, ICLR 2019.
Stars: ✭ 43 (+115%)
Word2VecfJavaWord2VecfJava: Java implementation of Dependency-Based Word Embeddings and extensions
Stars: ✭ 14 (-30%)
Dna2vecdna2vec: Consistent vector representations of variable-length k-mers
Stars: ✭ 117 (+485%)
wikidata-corpusTrain Wikidata with word2vec for word embedding tasks
Stars: ✭ 109 (+445%)
Active-Explainable-ClassificationA set of tools for leveraging pre-trained embeddings, active learning and model explainability for effecient document classification
Stars: ✭ 28 (+40%)
Spanish Word EmbeddingsSpanish word embeddings computed with different methods and from different corpora
Stars: ✭ 236 (+1080%)
Word-recognition-EmbedNet-CABCode implementation for our ICPR, 2020 paper titled "Improving Word Recognition using Multiple Hypotheses and Deep Embeddings"
Stars: ✭ 19 (-5%)
WordgcnACL 2019: Incorporating Syntactic and Semantic Information in Word Embeddings using Graph Convolutional Networks
Stars: ✭ 230 (+1050%)
Chameleon recsysSource code of CHAMELEON - A Deep Learning Meta-Architecture for News Recommender Systems
Stars: ✭ 202 (+910%)
JfasttextJava interface for fastText
Stars: ✭ 193 (+865%)
Vec4irWord Embeddings for Information Retrieval
Stars: ✭ 188 (+840%)
Naive-Resume-MatchingText Similarity Applied to resume, to compare Resumes with Job Descriptions and create a score to rank them. Similar to an ATS.
Stars: ✭ 27 (+35%)
TextheroText preprocessing, representation and visualization from zero to hero.
Stars: ✭ 2,407 (+11935%)
contextualLSTMContextual LSTM for NLP tasks like word prediction and word embedding creation for Deep Learning
Stars: ✭ 28 (+40%)
Sifrank zh基于预训练模型的中文关键词抽取方法(论文SIFRank: A New Baseline for Unsupervised Keyphrase Extraction Based on Pre-trained Language Model 的中文版代码)
Stars: ✭ 175 (+775%)
word2vec-tsneGoogle News and Leo Tolstoy: Visualizing Word2Vec Word Embeddings using t-SNE.
Stars: ✭ 59 (+195%)
GensimTopic Modelling for Humans
Stars: ✭ 12,763 (+63715%)
word2vec-on-wikipediaA pipeline for training word embeddings using word2vec on wikipedia corpus.
Stars: ✭ 68 (+240%)
PersianNERNamed-Entity Recognition in Persian Language
Stars: ✭ 48 (+140%)
Hash EmbeddingsPyTorch implementation of Hash Embeddings (NIPS 2017). Submission to the NIPS Implementation Challenge.
Stars: ✭ 126 (+530%)
SIFRankThe code of our paper "SIFRank: A New Baseline for Unsupervised Keyphrase Extraction Based on Pre-trained Language Model"
Stars: ✭ 96 (+380%)
wefeWEFE: The Word Embeddings Fairness Evaluation Framework. WEFE is a framework that standardizes the bias measurement and mitigation in Word Embeddings models. Please feel welcome to open an issue in case you have any questions or a pull request if you want to contribute to the project!
Stars: ✭ 164 (+720%)
codenamesCodenames AI using Word Vectors
Stars: ✭ 41 (+105%)
context2vecPyTorch implementation of context2vec from Melamud et al., CoNLL 2016
Stars: ✭ 18 (-10%)
datastories-semeval2017-task6Deep-learning model presented in "DataStories at SemEval-2017 Task 6: Siamese LSTM with Attention for Humorous Text Comparison".
Stars: ✭ 20 (+0%)
sisterSImple SenTence EmbeddeR
Stars: ✭ 66 (+230%)