All Projects → rusiaaman → PCPM

rusiaaman / PCPM

Licence: MIT license
Presenting Collection of Pretrained Models. Links to pretrained models in NLP and voice.

Projects that are alternatives of or similar to PCPM

Silero Models
Silero Models: pre-trained STT models and benchmarks made embarrassingly simple
Stars: ✭ 522 (+2385.71%)
Mutual labels:  speech-recognition, speech-to-text, pretrained-models, asr
Lingvo
Lingvo
Stars: ✭ 2,361 (+11142.86%)
Mutual labels:  speech-recognition, speech-to-text, language-model, asr
Asr audio data links
A list of publically available audio data that anyone can download for ASR or other speech activities
Stars: ✭ 128 (+509.52%)
Mutual labels:  speech-recognition, speech-to-text, asr
Speech To Text Russian
Проект для распознавания речи на русском языке на основе pykaldi.
Stars: ✭ 151 (+619.05%)
Mutual labels:  speech-recognition, speech-to-text, asr
speech-recognition-evaluation
Evaluate results from ASR/Speech-to-Text quickly
Stars: ✭ 25 (+19.05%)
Mutual labels:  speech-recognition, speech-to-text, asr
Mongolian Speech Recognition
Mongolian speech recognition with PyTorch
Stars: ✭ 97 (+361.9%)
Mutual labels:  speech-recognition, speech-to-text, asr
Vosk Api
Offline speech recognition API for Android, iOS, Raspberry Pi and servers with Python, Java, C# and Node
Stars: ✭ 1,357 (+6361.9%)
Mutual labels:  speech-recognition, speech-to-text, asr
Automatic Speech Recognition
🎧 Automatic Speech Recognition: DeepSpeech & Seq2Seq (TensorFlow)
Stars: ✭ 192 (+814.29%)
Mutual labels:  speech-recognition, speech-to-text, language-model
Pykaldi
A Python wrapper for Kaldi
Stars: ✭ 756 (+3500%)
Mutual labels:  speech-recognition, language-model, asr
megs
A merged version of multiple open-source German speech datasets.
Stars: ✭ 21 (+0%)
Mutual labels:  speech-recognition, speech-to-text, asr
Zeroth
Kaldi-based Korean ASR (한국어 음성인식) open-source project
Stars: ✭ 248 (+1080.95%)
Mutual labels:  speech-recognition, language-model, asr
leopard
On-device speech-to-text engine powered by deep learning
Stars: ✭ 354 (+1585.71%)
Mutual labels:  speech-recognition, speech-to-text, asr
Wav2letter
Speech Recognition model based off of FAIR research paper built using Pytorch.
Stars: ✭ 78 (+271.43%)
Mutual labels:  speech-recognition, speech-to-text, asr
Openasr
A pytorch based end2end speech recognition system.
Stars: ✭ 69 (+228.57%)
Mutual labels:  speech-recognition, speech-to-text, asr
Openseq2seq
Toolkit for efficient experimentation with Speech Recognition, Text2Speech and NLP
Stars: ✭ 1,378 (+6461.9%)
Mutual labels:  speech-recognition, speech-to-text, language-model
Syn Speech
Syn.Speech is a flexible speaker independent continuous speech recognition engine for Mono and .NET framework
Stars: ✭ 57 (+171.43%)
Mutual labels:  speech-recognition, speech-to-text, asr
ASR-Audio-Data-Links
A list of publically available audio data that anyone can download for ASR or other speech activities
Stars: ✭ 179 (+752.38%)
Mutual labels:  speech-recognition, speech-to-text, asr
Eesen
The official repository of the Eesen project
Stars: ✭ 738 (+3414.29%)
Mutual labels:  speech-recognition, speech-to-text, asr
Edgedict
Working online speech recognition based on RNN Transducer. ( Trained model release available in release )
Stars: ✭ 205 (+876.19%)
Mutual labels:  speech-recognition, speech-to-text, asr
wav2vec2-live
A live speech recognition using Facebooks wav2vec 2.0 model.
Stars: ✭ 205 (+876.19%)
Mutual labels:  speech-recognition, speech-to-text, asr

PCPM

Presenting Corpus of Pretrained Models. Links to pretrained models in NLP and voice with training script.

With rapid progress in NLP it is becoming easier to bootstrap a machine learning project involving text. Instead of starting with a base code, one can now start with a base pretrained model and within a few iterations get SOTA performance. This repository is made with the view that pretrained models minimizes collective human effort and cost of resources, thus accelerating development in the field.

Models listed are curated for either pytorch or tensorflow because of their wide usage.

Note: pytorch-transofmers is an awesome library which can be used to quickly infer/fine-tune from many pre-trained models in NLP. The pre-trained models from those are not included here.

Contents

Text ML

Language Models

Name Link Trained On Training script
Transformer-xl https://github.com/kimiyoung/transformer-xl/tree/master/tf#obtain-and-evaluate-pretrained-sota-models enwik8, lm1b, wt103, text8 https://github.com/kimiyoung/transformer-xl
GPT-2 https://github.com/openai/gpt-2/blob/master/download_model.py webtext https://github.com/nshepperd/gpt-2/
Adaptive Inputs (fairseq) https://github.com/pytorch/fairseq/blob/master/examples/language_model/README.md#pre-trained-models lm1b https://github.com/pytorch/fairseq/blob/master/examples/language_model/README.md

Permutation lanugage modelling Based - XLNet

Name Link Trained On Training script
XLnet https://github.com/zihangdai/xlnet/#released-models booksCorpus+English Wikipedia+Giga5+ClueWeb 2012-B+Common Crawl https://github.com/zihangdai/xlnet/

Masked Language Modelling Based - Bert

Name Link Trained On Training script
RoBERTa https://github.com/pytorch/fairseq/tree/master/examples/roberta#pre-trained-models booksCorpus+CC-N EWS+OpenWebText+CommonCrawl-Stories https://github.com/huggingface/transformers
BERT https://github.com/google-research/bert/ booksCorpus+English Wikipedia https://github.com/huggingface/transformers
MT-DNN https://mrc.blob.core.windows.net/mt-dnn-model/mt_dnn_base.pt (https://github.com/namisan/mt-dnn/blob/master/download.sh) glue https://github.com/namisan/mt-dnn

Machine Translation

Name Link Trained On Training script
OpenNMT http://opennmt.net/Models-py/ (pytorch) http://opennmt.net/Models-tf/ (tensorflow) English-German https://github.com/OpenNMT/OpenNMT-py
Fairseq (multiple models) https://github.com/pytorch/fairseq/blob/master/examples/translation/README.md#pre-trained-models WMT14 English-French, WMT16 English-German https://github.com/pytorch/fairseq/blob/master/examples/translation/README.md

Sentiment

Name Link Trained On Training script
Nvidia sentiment-discovery https://github.com/NVIDIA/sentiment-discovery#pretrained-models SST, imdb, Semeval-2018-tweet-emotion https://github.com/NVIDIA/sentiment-discovery
MT-DNN Sentiment https://drive.google.com/open?id=1-ld8_WpdQVDjPeYhb3AK8XYLGlZEbs-l SST https://github.com/namisan/mt-dnn

Reading Comprehension

SQUAD 1.1

Rank Name Link Training script
49 BiDaf https://s3-us-west-2.amazonaws.com/allennlp/models/bidaf-model-2017.09.15-charpad.tar.gz https://github.com/allenai/allennlp

Summarization

Model for English summarization

Name Link Trained On Training script
OpenNMT http://opennmt.net/Models-py/ Gigaword standard https://github.com/OpenNMT/OpenNMT-py

Speech to Text

Name Link Trained On Training script
NeMo-quartznet https://ngc.nvidia.com/catalog/models/nvidia:quartznet15x5 librispeech,mozilla-common-voice https://github.com/NVIDIA/NeMo
OpenSeq2Seq-Jasper https://nvidia.github.io/OpenSeq2Seq/html/speech-recognition.html#models librispeech https://github.com/NVIDIA/OpenSeq2Seq
Espnet https://github.com/espnet/espnet#asr-results librispeech,Aishell,HKUST,TEDLIUM2 https://github.com/espnet/espnet
wav2letter++ https://talonvoice.com/research/ librispeech https://github.com/facebookresearch/wav2letter
Deepspeech2 pytorch SeanNaren/deepspeech.pytorch#299 (comment) librispeech https://github.com/SeanNaren/deepspeech.pytorch
Deepspeech https://github.com/mozilla/DeepSpeech#getting-the-pre-trained-model mozilla-common-voice, librispeech, fisher, switchboard https://github.com/mozilla/DeepSpeech
speech-to-text-wavenet https://github.com/buriburisuri/speech-to-text-wavenet#pre-trained-models vctk https://github.com/buriburisuri/speech-to-text-wavenet
at16k https://github.com/at16k/at16k#download-models NA NA

Datasets

Datasets referenced in this document

Language Model data

Common crawl

http://commoncrawl.org/

enwik8

Wikipedia data dump (Large text compression benchmark) http://mattmahoney.net/dc/textdata.html

text8

Wikipedia cleaned text (Large text compression benchmark) http://mattmahoney.net/dc/textdata.html

lm1b

1 Billion Word Language Model Benchmark https://www.statmt.org/lm-benchmark/

wt103

Wikitext 103 https://blog.einstein.ai/the-wikitext-long-term-dependency-language-modeling-dataset/

webtext

Original dataset not released by the authors. An open source collection is available at https://skylion007.github.io/OpenWebTextCorpus/

English wikipedia

https://en.wikipedia.org/wiki/Wikipedia:Database_download#English-language_Wikipedia

BooksCorpus

https://yknzhu.wixsite.com/mbweb https://github.com/soskek/bookcorpus

Sentiment

SST

Stanford sentiment tree bank https://nlp.stanford.edu/sentiment/index.html. One of the Glue tasks.

IMDB

IMDB movie review dataset used for sentiment classification http://ai.stanford.edu/~amaas/data/sentiment

Semeval2018te

Semeval 2018 tweet emotion dataset https://competitions.codalab.org/competitions/17751

Glue

Glue is a collection of resources for benchmarking natural language systems. https://gluebenchmark.com/ Contains datasets on natural language inference, sentiment classification, paraphrase detection, similarity matching and lingusitc acceptability.

Speech to text data

fisher

https://pdfs.semanticscholar.org/a723/97679079439b075de815553c7b687ccfa886.pdf

librispeech

www.danielpovey.com/files/2015_icassp_librispeech.pdf

switchboard

https://ieeexplore.ieee.org/document/225858/

Mozilla common voice

https://github.com/mozilla/voice-web

vctk

https://datashare.is.ed.ac.uk/handle/10283/2651

Hall of Shame

High quality research which doesn't include pretrained models and/or code for public use.

Non English

Other Collections

Allen NLP

Built on pytorch, allen nlp has produced SOTA models and open sourced them. https://github.com/allenai/allennlp/blob/master/MODELS.md

They have neat interactive demo on various tasks at https://demo.allennlp.org/

GluonNLP

Based on MXNet this library has extensive list of pretrained models on various tasks in NLP. http://gluon-nlp.mxnet.io/master/index.html#model-zoo

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].