InltkNatural Language Toolkit for Indic Languages aims to provide out of the box support for various NLP tasks that an application developer might need
Stars: ✭ 702 (+197.46%)
FlairA very simple framework for state-of-the-art Natural Language Processing (NLP)
Stars: ✭ 11,065 (+4588.56%)
Textblob ArArabic support for textblob
Stars: ✭ 60 (-74.58%)
WordnetembeddingsObtaining word embeddings from a WordNet ontology
Stars: ✭ 33 (-86.02%)
DebiasweRemove problematic gender bias from word embeddings.
Stars: ✭ 175 (-25.85%)
Deep learning nlpKeras, PyTorch, and NumPy Implementations of Deep Learning Architectures for NLP
Stars: ✭ 407 (+72.46%)
MagnitudeA fast, efficient universal vector embedding utility package.
Stars: ✭ 1,394 (+490.68%)
NLP-StuffPrograms with word vectors, RNN, NLP stuff, etc
Stars: ✭ 19 (-91.95%)
Elmo TutorialA short tutorial on Elmo training (Pre trained, Training on new data, Incremental training)
Stars: ✭ 145 (-38.56%)
Lstm Context EmbeddingsAugmenting word embeddings with their surrounding context using bidirectional RNN
Stars: ✭ 57 (-75.85%)
Datastories Semeval2017 Task4Deep-learning model presented in "DataStories at SemEval-2017 Task 4: Deep LSTM with Attention for Message-level and Topic-based Sentiment Analysis".
Stars: ✭ 184 (-22.03%)
ScattertextBeautiful visualizations of how language differs among document types.
Stars: ✭ 1,722 (+629.66%)
Concise Ipython Notebooks For Deep LearningIpython Notebooks for solving problems like classification, segmentation, generation using latest Deep learning algorithms on different publicly available text and image data-sets.
Stars: ✭ 23 (-90.25%)
ShallowlearnAn experiment about re-implementing supervised learning models based on shallow neural network approaches (e.g. fastText) with some additional exclusive features and nice API. Written in Python and fully compatible with Scikit-learn.
Stars: ✭ 196 (-16.95%)
Nlp NotebooksA collection of notebooks for Natural Language Processing from NLP Town
Stars: ✭ 513 (+117.37%)
KadotKadot, the unsupervised natural language processing library.
Stars: ✭ 108 (-54.24%)
ChakinSimple downloader for pre-trained word vectors
Stars: ✭ 323 (+36.86%)
LftmImproving topic models LDA and DMM (one-topic-per-document model for short texts) with word embeddings (TACL 2015)
Stars: ✭ 168 (-28.81%)
neuralnets-semanticsWord semantics Deep Learning with Vanilla Python, Keras, Theano, TensorFlow, PyTorch
Stars: ✭ 15 (-93.64%)
Text SummarizerPython Framework for Extractive Text Summarization
Stars: ✭ 96 (-59.32%)
Glove As A Tensorflow Embedding LayerTaking a pretrained GloVe model, and using it as a TensorFlow embedding weight layer **inside the GPU**. Therefore, you only need to send the index of the words through the GPU data transfer bus, reducing data transfer overhead.
Stars: ✭ 85 (-63.98%)
Text-AnalysisExplaining textual analysis tools in Python. Including Preprocessing, Skip Gram (word2vec), and Topic Modelling.
Stars: ✭ 48 (-79.66%)
ClustercatFast Word Clustering Software
Stars: ✭ 65 (-72.46%)
Vec4irWord Embeddings for Information Retrieval
Stars: ✭ 188 (-20.34%)
Nlp overviewOverview of Modern Deep Learning Techniques Applied to Natural Language Processing
Stars: ✭ 1,104 (+367.8%)
Average Word2vec🔤 Calculate average word embeddings (word2vec) from documents for transfer learning
Stars: ✭ 52 (-77.97%)
Chameleon recsysSource code of CHAMELEON - A Deep Learning Meta-Architecture for News Recommender Systems
Stars: ✭ 202 (-14.41%)
EmbeddingsvizVisualize word embeddings of a vocabulary in TensorBoard, including the neighbors
Stars: ✭ 40 (-83.05%)
Hash EmbeddingsPyTorch implementation of Hash Embeddings (NIPS 2017). Submission to the NIPS Implementation Challenge.
Stars: ✭ 126 (-46.61%)
Top2vecTop2Vec learns jointly embedded topic, document and word vectors.
Stars: ✭ 972 (+311.86%)
TextheroText preprocessing, representation and visualization from zero to hero.
Stars: ✭ 2,407 (+919.92%)
Syntree2vecAn algorithm to augment syntactic hierarchy into word embeddings
Stars: ✭ 9 (-96.19%)
Dna2vecdna2vec: Consistent vector representations of variable-length k-mers
Stars: ✭ 117 (-50.42%)
Text2vecFast vectorization, topic modeling, distances and GloVe word embeddings in R.
Stars: ✭ 715 (+202.97%)
WordgcnACL 2019: Incorporating Syntactic and Semantic Information in Word Embeddings using Graph Convolutional Networks
Stars: ✭ 230 (-2.54%)
MetaA Modern C++ Data Sciences Toolkit
Stars: ✭ 600 (+154.24%)
DanlpDaNLP is a repository for Natural Language Processing resources for the Danish Language.
Stars: ✭ 111 (-52.97%)
Bert Embedding🔡 Token level embeddings from BERT model on mxnet and gluonnlp
Stars: ✭ 424 (+79.66%)
Sifrank zh基于预训练模型的中文关键词抽取方法(论文SIFRank: A New Baseline for Unsupervised Keyphrase Extraction Based on Pre-trained Language Model 的中文版代码)
Stars: ✭ 175 (-25.85%)
WegoWord Embeddings (e.g. Word2Vec) in Go!
Stars: ✭ 336 (+42.37%)
Easy BertA Dead Simple BERT API for Python and Java (https://github.com/google-research/bert)
Stars: ✭ 106 (-55.08%)
BiosentvecBioWordVec & BioSentVec: pre-trained embeddings for biomedical words and sentences
Stars: ✭ 308 (+30.51%)
JfasttextJava interface for fastText
Stars: ✭ 193 (-18.22%)
FastrtextR wrapper for fastText
Stars: ✭ 103 (-56.36%)
Lbl2VecLbl2Vec learns jointly embedded label, document and word vectors to retrieve documents with predefined topics from an unlabeled document corpus.
Stars: ✭ 25 (-89.41%)
GensimTopic Modelling for Humans
Stars: ✭ 12,763 (+5308.05%)
Postgres Word2vecutils to use word embedding like word2vec vectors in a postgres database
Stars: ✭ 96 (-59.32%)
KoanA word2vec negative sampling implementation with correct CBOW update.
Stars: ✭ 232 (-1.69%)
Question GenerationGenerating multiple choice questions from text using Machine Learning.
Stars: ✭ 227 (-3.81%)
GermanwordembeddingsToolkit to obtain and preprocess german corpora, train models using word2vec (gensim) and evaluate them with generated testsets
Stars: ✭ 189 (-19.92%)
MimickCode for Mimicking Word Embeddings using Subword RNNs (EMNLP 2017)
Stars: ✭ 152 (-35.59%)
Dict2vecDict2vec is a framework to learn word embeddings using lexical dictionaries.
Stars: ✭ 91 (-61.44%)