personality-predictionExperiments for automated personality detection using Language Models and psycholinguistic features on various famous personality datasets including the Essays dataset (Big-Five)
Stars: ✭ 109 (+303.7%)
pd3f🏭 PDF text extraction pipeline: self-hosted, local-first, Docker-based
Stars: ✭ 132 (+388.89%)
minGPT-TFA minimal TF2 re-implementation of the OpenAI GPT training
Stars: ✭ 36 (+33.33%)
CLSPCode and data for EMNLP 2018 paper "Cross-lingual Lexical Sememe Prediction"
Stars: ✭ 19 (-29.63%)
Pytorch NceThe Noise Contrastive Estimation for softmax output written in Pytorch
Stars: ✭ 204 (+655.56%)
gpt-j-apiAPI for the GPT-J language model 🦜. Including a FastAPI backend and a streamlit frontend
Stars: ✭ 248 (+818.52%)
KB-ALBERTKB국민은행에서 제공하는 경제/금융 도메인에 특화된 한국어 ALBERT 모델
Stars: ✭ 215 (+696.3%)
ZerothKaldi-based Korean ASR (한국어 음성인식) open-source project
Stars: ✭ 248 (+818.52%)
subword-lstm-lmLSTM Language Model with Subword Units Input Representations
Stars: ✭ 45 (+66.67%)
dasher-webDasher text entry in HTML, CSS, JavaScript, and SVG
Stars: ✭ 34 (+25.93%)
Gpt ScrollsA collaborative collection of open-source safe GPT-3 prompts that work well
Stars: ✭ 195 (+622.22%)
cscgCode Generation as a Dual Task of Code Summarization.
Stars: ✭ 28 (+3.7%)
lm-scorer📃Language Model based sentences scoring library
Stars: ✭ 264 (+877.78%)
FNet-pytorchUnofficial implementation of Google's FNet: Mixing Tokens with Fourier Transforms
Stars: ✭ 204 (+655.56%)
calmContext Aware Language Models
Stars: ✭ 29 (+7.41%)
gdcCode for the ICLR 2021 paper "A Distributional Approach to Controlled Text Generation"
Stars: ✭ 94 (+248.15%)
rnn-theanoRNN(LSTM, GRU) in Theano with mini-batch training; character-level language models in Theano
Stars: ✭ 68 (+151.85%)
tying-wv-and-wcImplementation for "Tying Word Vectors and Word Classifiers: A Loss Framework for Language Modeling"
Stars: ✭ 39 (+44.44%)
TF-NNLM-TKA toolkit for neural language modeling using Tensorflow including basic models like RNNs and LSTMs as well as more advanced models.
Stars: ✭ 20 (-25.93%)
PCPMPresenting Collection of Pretrained Models. Links to pretrained models in NLP and voice.
Stars: ✭ 21 (-22.22%)
Relational Rnn PytorchAn implementation of DeepMind's Relational Recurrent Neural Networks in PyTorch.
Stars: ✭ 236 (+774.07%)
Black-Box-TuningICML'2022: Black-Box Tuning for Language-Model-as-a-Service
Stars: ✭ 99 (+266.67%)
LingvoLingvo
Stars: ✭ 2,361 (+8644.44%)
backpropBackprop makes it simple to use, finetune, and deploy state-of-the-art ML models.
Stars: ✭ 229 (+748.15%)
mlp-gpt-jaxA GPT, made only of MLPs, in Jax
Stars: ✭ 53 (+96.3%)
Char Rnn ChineseMulti-layer Recurrent Neural Networks (LSTM, GRU, RNN) for character-level language models in Torch. Based on code of https://github.com/karpathy/char-rnn. Support Chinese and other things.
Stars: ✭ 192 (+611.11%)
mongolian-nlpUseful resources for Mongolian NLP
Stars: ✭ 119 (+340.74%)
Word-Prediction-NgramNext Word Prediction using n-gram Probabilistic Model with various Smoothing Techniques
Stars: ✭ 25 (-7.41%)
swig-srilmSWIG Wrapper for the SRILM toolkit
Stars: ✭ 33 (+22.22%)
CoLAKECOLING'2020: CoLAKE: Contextualized Language and Knowledge Embedding
Stars: ✭ 86 (+218.52%)
gap-text2sqlGAP-text2SQL: Learning Contextual Representations for Semantic Parsing with Generation-Augmented Pre-Training
Stars: ✭ 83 (+207.41%)
MinTLMinTL: Minimalist Transfer Learning for Task-Oriented Dialogue Systems
Stars: ✭ 61 (+125.93%)
CharLMCharacter-aware Neural Language Model implemented by PyTorch
Stars: ✭ 32 (+18.52%)
LM-CNLCChinese Natural Language Correction via Language Model
Stars: ✭ 15 (-44.44%)
asr2424-hour Automatic Speech Recognition
Stars: ✭ 27 (+0%)
language-plannerOfficial Code for "Language Models as Zero-Shot Planners: Extracting Actionable Knowledge for Embodied Agents"
Stars: ✭ 84 (+211.11%)
Vaaku2VecLanguage Modeling and Text Classification in Malayalam Language using ULMFiT
Stars: ✭ 68 (+151.85%)
Character-enhanced-Sememe-PredictionCode accompanying Incorporating Chinese Characters of Words for Lexical Sememe Prediction (ACL2018) https://arxiv.org/abs/1806.06349
Stars: ✭ 22 (-18.52%)
COCO-LM[NeurIPS 2021] COCO-LM: Correcting and Contrasting Text Sequences for Language Model Pretraining
Stars: ✭ 109 (+303.7%)
wechselCode for WECHSEL: Effective initialization of subword embeddings for cross-lingual transfer of monolingual language models.
Stars: ✭ 39 (+44.44%)
PLBARTOfficial code of our work, Unified Pre-training for Program Understanding and Generation [NAACL 2021].
Stars: ✭ 151 (+459.26%)
Mead BaselineDeep-Learning Model Exploration and Development for NLP
Stars: ✭ 238 (+781.48%)
Xlnet zh中文预训练XLNet模型: Pre-Trained Chinese XLNet_Large
Stars: ✭ 207 (+666.67%)
CodeT5Code for CodeT5: a new code-aware pre-trained encoder-decoder model.
Stars: ✭ 390 (+1344.44%)
Attention MechanismsImplementations for a family of attention mechanisms, suitable for all kinds of natural language processing tasks and compatible with TensorFlow 2.0 and Keras.
Stars: ✭ 203 (+651.85%)
mlmachine learning
Stars: ✭ 29 (+7.41%)
miniconsUtility for analyzing Transformer based representations of language.
Stars: ✭ 28 (+3.7%)
pyVHDLParserStreaming based VHDL parser.
Stars: ✭ 51 (+88.89%)
gpt-jA GPT-J API to use with python3 to generate text, blogs, code, and more
Stars: ✭ 101 (+274.07%)
open clipAn open source implementation of CLIP.
Stars: ✭ 1,534 (+5581.48%)
sememe predictionCodes for Lexical Sememe Prediction via Word Embeddings and Matrix Factorization (IJCAI 2017).
Stars: ✭ 59 (+118.52%)