Deep learning nlpKeras, PyTorch, and NumPy Implementations of Deep Learning Architectures for NLP
Stars: ✭ 407 (+273.39%)
lda2vecMixing Dirichlet Topic Models and Word Embeddings to Make lda2vec from this paper https://arxiv.org/abs/1605.02019
Stars: ✭ 27 (-75.23%)
codenamesCodenames AI using Word Vectors
Stars: ✭ 41 (-62.39%)
DebiasweRemove problematic gender bias from word embeddings.
Stars: ✭ 175 (+60.55%)
word2vec-on-wikipediaA pipeline for training word embeddings using word2vec on wikipedia corpus.
Stars: ✭ 68 (-37.61%)
Glove As A Tensorflow Embedding LayerTaking a pretrained GloVe model, and using it as a TensorFlow embedding weight layer **inside the GPU**. Therefore, you only need to send the index of the words through the GPU data transfer bus, reducing data transfer overhead.
Stars: ✭ 85 (-22.02%)
GensimTopic Modelling for Humans
Stars: ✭ 12,763 (+11609.17%)
Chameleon recsysSource code of CHAMELEON - A Deep Learning Meta-Architecture for News Recommender Systems
Stars: ✭ 202 (+85.32%)
word-benchmarksBenchmarks for intrinsic word embeddings evaluation.
Stars: ✭ 45 (-58.72%)
Dna2vecdna2vec: Consistent vector representations of variable-length k-mers
Stars: ✭ 117 (+7.34%)
WegoWord Embeddings (e.g. Word2Vec) in Go!
Stars: ✭ 336 (+208.26%)
KoanA word2vec negative sampling implementation with correct CBOW update.
Stars: ✭ 232 (+112.84%)
Postgres Word2vecutils to use word embedding like word2vec vectors in a postgres database
Stars: ✭ 96 (-11.93%)
word embeddingSample code for training Word2Vec and FastText using wiki corpus and their pretrained word embedding..
Stars: ✭ 21 (-80.73%)
word2vec-tsneGoogle News and Leo Tolstoy: Visualizing Word2Vec Word Embeddings using t-SNE.
Stars: ✭ 59 (-45.87%)
Text-AnalysisExplaining textual analysis tools in Python. Including Preprocessing, Skip Gram (word2vec), and Topic Modelling.
Stars: ✭ 48 (-55.96%)
NTUA-slp-nlp💻Speech and Natural Language Processing (SLP & NLP) Lab Assignments for ECE NTUA
Stars: ✭ 19 (-82.57%)
Text SummarizerPython Framework for Extractive Text Summarization
Stars: ✭ 96 (-11.93%)
MagnitudeA fast, efficient universal vector embedding utility package.
Stars: ✭ 1,394 (+1178.9%)
Dict2vecDict2vec is a framework to learn word embeddings using lexical dictionaries.
Stars: ✭ 91 (-16.51%)
GermanwordembeddingsToolkit to obtain and preprocess german corpora, train models using word2vec (gensim) and evaluate them with generated testsets
Stars: ✭ 189 (+73.39%)
ShallowlearnAn experiment about re-implementing supervised learning models based on shallow neural network approaches (e.g. fastText) with some additional exclusive features and nice API. Written in Python and fully compatible with Scikit-learn.
Stars: ✭ 196 (+79.82%)
Text2vecFast vectorization, topic modeling, distances and GloVe word embeddings in R.
Stars: ✭ 715 (+555.96%)
SWDMSIGIR 2017: Embedding-based query expansion for weighted sequential dependence retrieval model
Stars: ✭ 35 (-67.89%)
ScattertextBeautiful visualizations of how language differs among document types.
Stars: ✭ 1,722 (+1479.82%)
Word2vec訓練中文詞向量 Word2vec, Word2vec was created by a team of researchers led by Tomas Mikolov at Google.
Stars: ✭ 48 (-55.96%)
two-stream-cnnA two-stream convolutional neural network for learning abitrary similarity functions over two sets of training data
Stars: ✭ 24 (-77.98%)
word2vecUse word2vec to improve search result
Stars: ✭ 63 (-42.2%)
DeepLearning-LabCode lab for deep learning. Including rnn,seq2seq,word2vec,cross entropy,bidirectional rnn,convolution operation,pooling operation,InceptionV3,transfer learning.
Stars: ✭ 83 (-23.85%)
Word2VecJavaWord2Vec In Java (2013 google word2vec opensource)
Stars: ✭ 13 (-88.07%)
entity-fishingA machine learning tool for fishing entities
Stars: ✭ 176 (+61.47%)
SentimentAnalysis(BOW, TF-IDF, Word2Vec, BERT) Word Embeddings + (SVM, Naive Bayes, Decision Tree, Random Forest) Base Classifiers + Pre-trained BERT on Tensorflow Hub + 1-D CNN and Bi-Directional LSTM on IMDB Movie Reviews Dataset
Stars: ✭ 40 (-63.3%)
SIFRankThe code of our paper "SIFRank: A New Baseline for Unsupervised Keyphrase Extraction Based on Pre-trained Language Model"
Stars: ✭ 96 (-11.93%)
NLP PEMDCNLP Predtrained Embeddings, Models and Datasets Collections(NLP_PEMDC). The collection will keep updating.
Stars: ✭ 58 (-46.79%)
wmd4jwmd4j is a Java library for calculating Word Mover's Distance (WMD)
Stars: ✭ 31 (-71.56%)
doubanIMDbIMDb + Rotten Tomatoes + Wikipedia on Douban Movie
Stars: ✭ 93 (-14.68%)
QuestionClusteringClasificador de preguntas escrito en python 3 que fue implementado en el siguiente vídeo: https://youtu.be/qnlW1m6lPoY
Stars: ✭ 15 (-86.24%)
RolXAn alternative implementation of Recursive Feature and Role Extraction (KDD11 & KDD12)
Stars: ✭ 52 (-52.29%)
dnn-lstm-word-segmentChinese Word Segmention Base on the Deep Learning and LSTM Neural Network
Stars: ✭ 24 (-77.98%)
test word2vec uyghurBu Uyghur yéziqini Pythonning gensim ambiridiki word2vec algorizimida sinap baqqan misal.
Stars: ✭ 15 (-86.24%)
word2vecRust interface to word2vec.
Stars: ✭ 22 (-79.82%)
wikiapiJavaScript MediaWiki API for node.js
Stars: ✭ 28 (-74.31%)
compress-fasttextTools for shrinking fastText models (in gensim format)
Stars: ✭ 124 (+13.76%)
walkletsA lightweight implementation of Walklets from "Don't Walk Skip! Online Learning of Multi-scale Network Embeddings" (ASONAM 2017).
Stars: ✭ 94 (-13.76%)
transparencia-dados-abertos-brasilA survey of Brazilian states' and municipalities' transparency and open data portals, as well as institutional websites, obtained from several public data sources. 🇧🇷 Levantamento de portais estaduais e municipais de transparência e dados abertos, bem como os portais institucionais, obtido a partir de diversas fontes públicas de dados.
Stars: ✭ 46 (-57.8%)
ordiaWikidata lexemes presentations
Stars: ✭ 21 (-80.73%)
wdumperTool for generating filtered Wikidata RDF exports
Stars: ✭ 25 (-77.06%)
reveryA personal semantic search engine capable of surfacing relevant bookmarks, journal entries, notes, blogs, contacts, and more, built on an efficient document embedding algorithm and Monocle's personal search index.
Stars: ✭ 200 (+83.49%)
word2vec pipelineNLP pipeline using word2vec (preprocessing/embedding/prediction/clustering)
Stars: ✭ 108 (-0.92%)
biovecProtVec can be used in protein interaction predictions, structure prediction, and protein data visualization.
Stars: ✭ 23 (-78.9%)