Transformers🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
Stars: ✭ 55,742 (+105073.58%)
pd3f🏭 PDF text extraction pipeline: self-hosted, local-first, Docker-based
Stars: ✭ 132 (+149.06%)
Nlp learning结合python一起学习自然语言处理 (nlp): 语言模型、HMM、PCFG、Word2vec、完形填空式阅读理解任务、朴素贝叶斯分类器、TFIDF、PCA、SVD
Stars: ✭ 188 (+254.72%)
Xlnet GenXLNet for generating language.
Stars: ✭ 164 (+209.43%)
digitRecognitionImplementation of a digit recognition using my Neural Network with the MNIST data set.
Stars: ✭ 21 (-60.38%)
OptimusOptimus: the first large-scale pre-trained VAE language model
Stars: ✭ 180 (+239.62%)
CharLMCharacter-aware Neural Language Model implemented by PyTorch
Stars: ✭ 32 (-39.62%)
PLBARTOfficial code of our work, Unified Pre-training for Program Understanding and Generation [NAACL 2021].
Stars: ✭ 151 (+184.91%)
F LmLanguage Modeling
Stars: ✭ 156 (+194.34%)
Attention MechanismsImplementations for a family of attention mechanisms, suitable for all kinds of natural language processing tasks and compatible with TensorFlow 2.0 and Keras.
Stars: ✭ 203 (+283.02%)
KB-ALBERTKB국민은행에서 제공하는 경제/금융 도메인에 특화된 한국어 ALBERT 모델
Stars: ✭ 215 (+305.66%)
Char Rnn ChineseMulti-layer Recurrent Neural Networks (LSTM, GRU, RNN) for character-level language models in Torch. Based on code of https://github.com/karpathy/char-rnn. Support Chinese and other things.
Stars: ✭ 192 (+262.26%)
jax-cfdComputational Fluid Dynamics in JAX
Stars: ✭ 399 (+652.83%)
Bert Sklearna sklearn wrapper for Google's BERT model
Stars: ✭ 182 (+243.4%)
rnn-theanoRNN(LSTM, GRU) in Theano with mini-batch training; character-level language models in Theano
Stars: ✭ 68 (+28.3%)
Gpt NeoAn implementation of model parallel GPT2& GPT3-like models, with the ability to scale up to full GPT3 sizes (and possibly more!), using the mesh-tensorflow library.
Stars: ✭ 1,252 (+2262.26%)
gap-text2sqlGAP-text2SQL: Learning Contextual Representations for Semantic Parsing with Generation-Augmented Pre-Training
Stars: ✭ 83 (+56.6%)
Lotclass[EMNLP 2020] Text Classification Using Label Names Only: A Language Model Self-Training Approach
Stars: ✭ 160 (+201.89%)
TF-NNLM-TKA toolkit for neural language modeling using Tensorflow including basic models like RNNs and LSTMs as well as more advanced models.
Stars: ✭ 20 (-62.26%)
SpeechtAn opensource speech-to-text software written in tensorflow
Stars: ✭ 152 (+186.79%)
jax-rlJAX implementations of core Deep RL algorithms
Stars: ✭ 61 (+15.09%)
Mead BaselineDeep-Learning Model Exploration and Development for NLP
Stars: ✭ 238 (+349.06%)
TupeTransformer with Untied Positional Encoding (TUPE). Code of paper "Rethinking Positional Encoding in Language Pre-training". Improve existing models like BERT.
Stars: ✭ 143 (+169.81%)
Pytorch NceThe Noise Contrastive Estimation for softmax output written in Pytorch
Stars: ✭ 204 (+284.91%)
asr2424-hour Automatic Speech Recognition
Stars: ✭ 27 (-49.06%)
LingvoLingvo
Stars: ✭ 2,361 (+4354.72%)
jax-modelsUnofficial JAX implementations of deep learning research papers
Stars: ✭ 108 (+103.77%)
Gpt ScrollsA collaborative collection of open-source safe GPT-3 prompts that work well
Stars: ✭ 195 (+267.92%)
omdJAX code for the paper "Control-Oriented Model-Based Reinforcement Learning with Implicit Differentiation"
Stars: ✭ 43 (-18.87%)
lm-scorer📃Language Model based sentences scoring library
Stars: ✭ 264 (+398.11%)
Bert As Language Modelbert as language model, fork from https://github.com/google-research/bert
Stars: ✭ 185 (+249.06%)
Vaaku2VecLanguage Modeling and Text Classification in Malayalam Language using ULMFiT
Stars: ✭ 68 (+28.3%)
Keras BertImplementation of BERT that could load official pre-trained models for feature extraction and prediction
Stars: ✭ 2,264 (+4171.7%)
personality-predictionExperiments for automated personality detection using Language Models and psycholinguistic features on various famous personality datasets including the Essays dataset (Big-Five)
Stars: ✭ 109 (+105.66%)
MacbertRevisiting Pre-trained Models for Chinese Natural Language Processing (Findings of EMNLP)
Stars: ✭ 167 (+215.09%)
MLPA multilayer perceptron in JavaScript
Stars: ✭ 15 (-71.7%)
Indic BertBERT-based Multilingual Model for Indian Languages
Stars: ✭ 160 (+201.89%)
LazynlpLibrary to scrape and clean web pages to create massive datasets.
Stars: ✭ 1,985 (+3645.28%)
COCO-LM[NeurIPS 2021] COCO-LM: Correcting and Contrasting Text Sequences for Language Model Pretraining
Stars: ✭ 109 (+105.66%)
Keras XlnetImplementation of XLNet that can load pretrained checkpoints
Stars: ✭ 159 (+200%)
calmContext Aware Language Models
Stars: ✭ 29 (-45.28%)
Transformer LmTransformer language model (GPT-2) with sentencepiece tokenizer
Stars: ✭ 154 (+190.57%)
graphsignalGraphsignal Python agent
Stars: ✭ 158 (+198.11%)
Electra pytorchPretrain and finetune ELECTRA with fastai and huggingface. (Results of the paper replicated !)
Stars: ✭ 149 (+181.13%)
g-mlp-pytorchImplementation of gMLP, an all-MLP replacement for Transformers, in Pytorch
Stars: ✭ 383 (+622.64%)
Awd Lstm LmLSTM and QRNN Language Model Toolkit for PyTorch
Stars: ✭ 1,834 (+3360.38%)
ZerothKaldi-based Korean ASR (한국어 음성인식) open-source project
Stars: ✭ 248 (+367.92%)
Ld NetEfficient Contextualized Representation: Language Model Pruning for Sequence Labeling
Stars: ✭ 148 (+179.25%)
Clue中文语言理解测评基准 Chinese Language Understanding Evaluation Benchmark: datasets, baselines, pre-trained models, corpus and leaderboard
Stars: ✭ 2,425 (+4475.47%)
Relational Rnn PytorchAn implementation of DeepMind's Relational Recurrent Neural Networks in PyTorch.
Stars: ✭ 236 (+345.28%)
dasher-webDasher text entry in HTML, CSS, JavaScript, and SVG
Stars: ✭ 34 (-35.85%)
swig-srilmSWIG Wrapper for the SRILM toolkit
Stars: ✭ 33 (-37.74%)
ADAMADAM implements a collection of algorithms for calculating rigid-body dynamics in Jax, CasADi, PyTorch, and Numpy.
Stars: ✭ 51 (-3.77%)