Chinese ElectraPre-trained Chinese ELECTRA(中文ELECTRA预训练模型)
Stars: ✭ 830 (-63.34%)
GetlangNatural language detection package in pure Go
Stars: ✭ 110 (-95.14%)
Xlnet GenXLNet for generating language.
Stars: ✭ 164 (-92.76%)
TnerLanguage model finetuning on NER with an easy interface, and cross-domain evaluation. We released NER models finetuned on various domain via huggingface model hub.
Stars: ✭ 54 (-97.61%)
kwxBERT, LDA, and TFIDF based keyword extraction in Python
Stars: ✭ 33 (-98.54%)
ChineseglueLanguage Understanding Evaluation benchmark for Chinese: datasets, baselines, pre-trained models,corpus and leaderboard
Stars: ✭ 1,548 (-31.63%)
Dl Nlp ReadingsMy Reading Lists of Deep Learning and Natural Language Processing
Stars: ✭ 656 (-71.02%)
Awesome Bert NlpA curated list of NLP resources focused on BERT, attention mechanism, Transformer networks, and transfer learning.
Stars: ✭ 567 (-74.96%)
Openseq2seqToolkit for efficient experimentation with Speech Recognition, Text2Speech and NLP
Stars: ✭ 1,378 (-39.13%)
Spark NlpState of the Art Natural Language Processing
Stars: ✭ 2,518 (+11.22%)
Albert pytorchA Lite Bert For Self-Supervised Learning Language Representations
Stars: ✭ 539 (-76.19%)
Bert As ServiceMapping a variable-length sentence to a fixed-length vector using BERT model
Stars: ✭ 9,779 (+331.93%)
TupeTransformer with Untied Positional Encoding (TUPE). Code of paper "Rethinking Positional Encoding in Language Pre-training". Improve existing models like BERT.
Stars: ✭ 143 (-93.68%)
KoBERT-nsmcNaver movie review sentiment classification with KoBERT
Stars: ✭ 57 (-97.48%)
Fast BertSuper easy library for BERT based NLP models
Stars: ✭ 1,678 (-25.88%)
LmchallengeA library & tools to evaluate predictive language models.
Stars: ✭ 47 (-97.92%)
knowledge-graph-nlp-in-action从模型训练到部署,实战知识图谱(Knowledge Graph)&自然语言处理(NLP)。涉及 Tensorflow, Bert+Bi-LSTM+CRF,Neo4j等 涵盖 Named Entity Recognition,Text Classify,Information Extraction,Relation Extraction 等任务。
Stars: ✭ 58 (-97.44%)
TongramsA C++ library providing fast language model queries in compressed space.
Stars: ✭ 88 (-96.11%)
Lotclass[EMNLP 2020] Text Classification Using Label Names Only: A Language Model Self-Training Approach
Stars: ✭ 160 (-92.93%)
Zamia SpeechOpen tools and data for cloudless automatic speech recognition
Stars: ✭ 374 (-83.48%)
Bit RnnQuantize weights and activations in Recurrent Neural Networks.
Stars: ✭ 86 (-96.2%)
Kogpt2Korean GPT-2 pretrained cased (KoGPT2)
Stars: ✭ 368 (-83.75%)
Mt DnnMulti-Task Deep Neural Networks for Natural Language Understanding
Stars: ✭ 1,871 (-17.36%)
Azureml BertEnd-to-End recipes for pre-training and fine-tuning BERT using Azure Machine Learning Service
Stars: ✭ 342 (-84.89%)
Bio embeddingsGet protein embeddings from protein sequences
Stars: ✭ 86 (-96.2%)
TrankitTrankit is a Light-Weight Transformer-based Python Toolkit for Multilingual Natural Language Processing
Stars: ✭ 311 (-86.26%)
Pycorrectorpycorrector is a toolkit for text error correction. 文本纠错,Kenlm,Seq2Seq_Attention,BERT,MacBERT,ELECTRA,ERNIE,Transformer等模型实现,开箱即用。
Stars: ✭ 2,857 (+26.19%)
Xlnet PytorchAn implementation of Google Brain's 2019 XLNet in PyTorch
Stars: ✭ 304 (-86.57%)
Nlp TutorialNatural Language Processing Tutorial for Deep Learning Researchers
Stars: ✭ 9,895 (+337.06%)
BertweetBERTweet: A pre-trained language model for English Tweets (EMNLP-2020)
Stars: ✭ 282 (-87.54%)
Awesome Bertbert nlp papers, applications and github resources, including the newst xlnet , BERT、XLNet 相关论文和 github 项目
Stars: ✭ 1,732 (-23.5%)
Nezha chinese pytorchNEZHA: Neural Contextualized Representation for Chinese Language Understanding
Stars: ✭ 65 (-97.13%)
FewCLUEFewCLUE 小样本学习测评基准,中文版
Stars: ✭ 251 (-88.91%)
F LmLanguage Modeling
Stars: ✭ 156 (-93.11%)
few-shot-lmThe source code of "Language Models are Few-shot Multilingual Learners" (MRL @ EMNLP 2021)
Stars: ✭ 32 (-98.59%)
PIEFast + Non-Autoregressive Grammatical Error Correction using BERT. Code and Pre-trained models for paper "Parallel Iterative Edit Models for Local Sequence Transduction": www.aclweb.org/anthology/D19-1435.pdf (EMNLP-IJCNLP 2019)
Stars: ✭ 164 (-92.76%)
Chars2vecCharacter-based word embeddings model based on RNN for handling real world texts
Stars: ✭ 130 (-94.26%)
CLUE pytorchCLUE baseline pytorch CLUE的pytorch版本基线
Stars: ✭ 72 (-96.82%)
Char rnn lm zhlanguage model in Chinese,基于Pytorch官方文档实现
Stars: ✭ 57 (-97.48%)
python-arpa🐍 Python library for n-gram models in ARPA format
Stars: ✭ 35 (-98.45%)
Gpt NeoAn implementation of model parallel GPT2& GPT3-like models, with the ability to scale up to full GPT3 sizes (and possibly more!), using the mesh-tensorflow library.
Stars: ✭ 1,252 (-44.7%)
bert-as-a-service TFXEnd-to-end pipeline with TFX to train and deploy a BERT model for sentiment analysis.
Stars: ✭ 32 (-98.59%)
keras-bert-nerKeras solution of Chinese NER task using BiLSTM-CRF/BiGRU-CRF/IDCNN-CRF model with Pretrained Language Model: supporting BERT/RoBERTa/ALBERT
Stars: ✭ 7 (-99.69%)
policy-data-analyzerBuilding a model to recognize incentives for landscape restoration in environmental policies from Latin America, the US and India. Bringing NLP to the world of policy analysis through an extensible framework that includes scraping, preprocessing, active learning and text analysis pipelines.
Stars: ✭ 22 (-99.03%)
SuggestTop-k Approximate String Matching.
Stars: ✭ 50 (-97.79%)
VideoBERTUsing VideoBERT to tackle video prediction
Stars: ✭ 56 (-97.53%)
SpeechtAn opensource speech-to-text software written in tensorflow
Stars: ✭ 152 (-93.29%)
ADL2019Applied Deep Learning (2019 Spring) @ NTU
Stars: ✭ 20 (-99.12%)
RobbertA Dutch RoBERTa-based language model
Stars: ✭ 120 (-94.7%)
Nlp Librarycurated collection of papers for the nlp practitioner 📖👩🔬
Stars: ✭ 1,025 (-54.73%)
Gpt2 FrenchGPT-2 French demo | Démo française de GPT-2
Stars: ✭ 47 (-97.92%)
OptimusOptimus: the first large-scale pre-trained VAE language model
Stars: ✭ 180 (-92.05%)
KashgariKashgari is a production-level NLP Transfer learning framework built on top of tf.keras for text-labeling and text-classification, includes Word2Vec, BERT, and GPT2 Language Embedding.
Stars: ✭ 2,235 (-1.28%)