Xlnet GenXLNet for generating language.
Stars: ✭ 164 (-20.77%)
RobbertA Dutch RoBERTa-based language model
Stars: ✭ 120 (-42.03%)
Lm Lstm CrfEmpower Sequence Labeling with Task-Aware Language Model
Stars: ✭ 778 (+275.85%)
Nlp learning结合python一起学习自然语言处理 (nlp): 语言模型、HMM、PCFG、Word2vec、完形填空式阅读理解任务、朴素贝叶斯分类器、TFIDF、PCA、SVD
Stars: ✭ 188 (-9.18%)
Lightnlp基于Pytorch和torchtext的自然语言处理深度学习框架。
Stars: ✭ 739 (+257%)
Lingopackage lingo provides the data structures and algorithms required for natural language processing
Stars: ✭ 113 (-45.41%)
Lotclass[EMNLP 2020] Text Classification Using Label Names Only: A Language Model Self-Training Approach
Stars: ✭ 160 (-22.71%)
KobertKorean BERT pre-trained cased (KoBERT)
Stars: ✭ 591 (+185.51%)
GetlangNatural language detection package in pure Go
Stars: ✭ 110 (-46.86%)
Albert pytorchA Lite Bert For Self-Supervised Learning Language Representations
Stars: ✭ 539 (+160.39%)
Easy BertA Dead Simple BERT API for Python and Java (https://github.com/google-research/bert)
Stars: ✭ 106 (-48.79%)
F LmLanguage Modeling
Stars: ✭ 156 (-24.64%)
Bert PytorchGoogle AI 2018 BERT pytorch implementation
Stars: ✭ 4,642 (+2142.51%)
Pytorch gbw lmPyTorch Language Model for 1-Billion Word (LM1B / GBW) Dataset
Stars: ✭ 101 (-51.21%)
CtcwordbeamsearchConnectionist Temporal Classification (CTC) decoder with dictionary and language model for TensorFlow.
Stars: ✭ 398 (+92.27%)
Bert Sklearna sklearn wrapper for Google's BERT model
Stars: ✭ 182 (-12.08%)
Tf chatbot seq2seq antilmSeq2seq chatbot with attention and anti-language model to suppress generic response, option for further improve by deep reinforcement learning.
Stars: ✭ 369 (+78.26%)
TongramsA C++ library providing fast language model queries in compressed space.
Stars: ✭ 88 (-57.49%)
Azureml BertEnd-to-End recipes for pre-training and fine-tuning BERT using Azure Machine Learning Service
Stars: ✭ 342 (+65.22%)
SpeechtAn opensource speech-to-text software written in tensorflow
Stars: ✭ 152 (-26.57%)
Gpt NeoxAn implementation of model parallel GPT-3-like models on GPUs, based on the DeepSpeed library. Designed to be able to train models in the hundreds of billions of parameters or larger.
Stars: ✭ 303 (+46.38%)
Bit RnnQuantize weights and activations in Recurrent Neural Networks.
Stars: ✭ 86 (-58.45%)
Transfer NlpNLP library designed for reproducible experimentation management
Stars: ✭ 287 (+38.65%)
Attention MechanismsImplementations for a family of attention mechanisms, suitable for all kinds of natural language processing tasks and compatible with TensorFlow 2.0 and Keras.
Stars: ✭ 203 (-1.93%)
BluebertBlueBERT, pre-trained on PubMed abstracts and clinical notes (MIMIC-III).
Stars: ✭ 273 (+31.88%)
Bio embeddingsGet protein embeddings from protein sequences
Stars: ✭ 86 (-58.45%)
few-shot-lmThe source code of "Language Models are Few-shot Multilingual Learners" (MRL @ EMNLP 2021)
Stars: ✭ 32 (-84.54%)
python-arpa🐍 Python library for n-gram models in ARPA format
Stars: ✭ 35 (-83.09%)
Full stack transformerPytorch library for end-to-end transformer models training, inference and serving
Stars: ✭ 71 (-65.7%)
SDLM-pytorchCode accompanying EMNLP 2018 paper Language Modeling with Sparse Product of Sememe Experts
Stars: ✭ 27 (-86.96%)
OptimusOptimus: the first large-scale pre-trained VAE language model
Stars: ✭ 180 (-13.04%)
Cross Domain nerCross-domain NER using cross-domain language modeling, code for ACL 2019 paper
Stars: ✭ 67 (-67.63%)
MinTLMinTL: Minimalist Transfer Learning for Task-Oriented Dialogue Systems
Stars: ✭ 61 (-70.53%)
CodeT5Code for CodeT5: a new code-aware pre-trained encoder-decoder model.
Stars: ✭ 390 (+88.41%)
Gpt2PyTorch Implementation of OpenAI GPT-2
Stars: ✭ 64 (-69.08%)
Word-Prediction-NgramNext Word Prediction using n-gram Probabilistic Model with various Smoothing Techniques
Stars: ✭ 25 (-87.92%)
Char Rnn ChineseMulti-layer Recurrent Neural Networks (LSTM, GRU, RNN) for character-level language models in Torch. Based on code of https://github.com/karpathy/char-rnn. Support Chinese and other things.
Stars: ✭ 192 (-7.25%)
language-plannerOfficial Code for "Language Models as Zero-Shot Planners: Extracting Actionable Knowledge for Embodied Agents"
Stars: ✭ 84 (-59.42%)
PhonlpPhoNLP: A BERT-based multi-task learning toolkit for part-of-speech tagging, named entity recognition and dependency parsing (NAACL 2021)
Stars: ✭ 56 (-72.95%)
TupeTransformer with Untied Positional Encoding (TUPE). Code of paper "Rethinking Positional Encoding in Language Pre-training". Improve existing models like BERT.
Stars: ✭ 143 (-30.92%)
TnerLanguage model finetuning on NER with an easy interface, and cross-domain evaluation. We released NER models finetuned on various domain via huggingface model hub.
Stars: ✭ 54 (-73.91%)
mongolian-nlpUseful resources for Mongolian NLP
Stars: ✭ 119 (-42.51%)
Gpt NeoAn implementation of model parallel GPT2& GPT3-like models, with the ability to scale up to full GPT3 sizes (and possibly more!), using the mesh-tensorflow library.
Stars: ✭ 1,252 (+504.83%)
LmchallengeA library & tools to evaluate predictive language models.
Stars: ✭ 47 (-77.29%)
Pytorch NceThe Noise Contrastive Estimation for softmax output written in Pytorch
Stars: ✭ 204 (-1.45%)
LingvoLingvo
Stars: ✭ 2,361 (+1040.58%)
Indic BertBERT-based Multilingual Model for Indian Languages
Stars: ✭ 160 (-22.71%)
Chars2vecCharacter-based word embeddings model based on RNN for handling real world texts
Stars: ✭ 130 (-37.2%)
Pytorch CppC++ Implementation of PyTorch Tutorials for Everyone
Stars: ✭ 1,014 (+389.86%)