All Projects β†’ loretoparisi β†’ hf-experiments

loretoparisi / hf-experiments

Licence: MIT license
Experiments with Hugging Face πŸ”¬ πŸ€—

Programming Languages

python
139335 projects - #7 most used programming language
Jupyter Notebook
11667 projects
Cuda
1817 projects
C++
36643 projects - #6 most used programming language
HTML
75241 projects
shell
77523 projects
Dockerfile
14818 projects

Projects that are alternatives of or similar to hf-experiments

soxan
Wav2Vec for speech recognition, classification, and audio classification
Stars: ✭ 113 (+205.41%)
Mutual labels:  speech-recognition, automatic-speech-recognition
teanaps
μžμ—°μ–΄ μ²˜λ¦¬μ™€ ν…μŠ€νŠΈ 뢄석을 μœ„ν•œ μ˜€ν”ˆμ†ŒμŠ€ 파이썬 라이브러리 μž…λ‹ˆλ‹€.
Stars: ✭ 91 (+145.95%)
Mutual labels:  topic-modeling, summarization
sova-asr
SOVA ASR (Automatic Speech Recognition)
Stars: ✭ 123 (+232.43%)
Mutual labels:  speech-recognition, automatic-speech-recognition
wenet
Production First and Production Ready End-to-End Speech Recognition Toolkit
Stars: ✭ 2,384 (+6343.24%)
Mutual labels:  speech-recognition, automatic-speech-recognition
DrFAQ
DrFAQ is a plug-and-play question answering NLP chatbot that can be generally applied to any organisation's text corpora.
Stars: ✭ 29 (-21.62%)
Mutual labels:  question-answering, huggingface
kaldi-long-audio-alignment
Long audio alignment using Kaldi
Stars: ✭ 21 (-43.24%)
Mutual labels:  speech-recognition, automatic-speech-recognition
Awesome Speech Recognition Speech Synthesis Papers
Automatic Speech Recognition (ASR), Speaker Verification, Speech Synthesis, Text-to-Speech (TTS), Language Modelling, Singing Voice Synthesis (SVS), Voice Conversion (VC)
Stars: ✭ 2,085 (+5535.14%)
Mutual labels:  speech-recognition, automatic-speech-recognition
Paper Reading
Paper reading list in natural language processing, including dialogue systems and text generation related topics.
Stars: ✭ 508 (+1272.97%)
Mutual labels:  question-answering, topic-modeling
Scattertext
Beautiful visualizations of how language differs among document types.
Stars: ✭ 1,722 (+4554.05%)
Mutual labels:  sentiment, topic-modeling
PCPM
Presenting Collection of Pretrained Models. Links to pretrained models in NLP and voice.
Stars: ✭ 21 (-43.24%)
Mutual labels:  sentiment, speech-recognition
ml-with-audio
HF's ML for Audio study group
Stars: ✭ 104 (+181.08%)
Mutual labels:  speech-recognition, huggingface
obvi
A Polymer 3+ webcomponent / button for doing speech recognition
Stars: ✭ 54 (+45.95%)
Mutual labels:  speech-recognition, automatic-speech-recognition
2018-dlsl
UPC Deep Learning for Speech and Language 2018
Stars: ✭ 18 (-51.35%)
Mutual labels:  speech-recognition, automatic-speech-recognition
deep avsr
A PyTorch implementation of the Deep Audio-Visual Speech Recognition paper.
Stars: ✭ 104 (+181.08%)
Mutual labels:  speech-recognition, automatic-speech-recognition
Haystack
πŸ” Haystack is an open source NLP framework that leverages Transformer models. It enables developers to implement production-ready neural search, question answering, semantic document search and summarization for a wide range of applications.
Stars: ✭ 3,409 (+9113.51%)
Mutual labels:  question-answering, summarization
demo vietasr
Vietnamese Speech Recognition
Stars: ✭ 22 (-40.54%)
Mutual labels:  speech-recognition, automatic-speech-recognition
query-focused-sum
Official code repository for "Exploring Neural Models for Query-Focused Summarization".
Stars: ✭ 17 (-54.05%)
Mutual labels:  question-answering, summarization
text2text
Text2Text: Cross-lingual natural language processing and generation toolkit
Stars: ✭ 188 (+408.11%)
Mutual labels:  question-answering, summarization
Automatic speech recognition
End-to-end Automatic Speech Recognition for Madarian and English in Tensorflow
Stars: ✭ 2,751 (+7335.14%)
Mutual labels:  speech-recognition, automatic-speech-recognition
UHV-OTS-Speech
A data annotation pipeline to generate high-quality, large-scale speech datasets with machine pre-labeling and fully manual auditing.
Stars: ✭ 94 (+154.05%)
Mutual labels:  speech-recognition, topic-detection

hf-experiments

Machine Learning (cool) Experiments πŸ”¬ πŸ€— with Hugging Face's (HF) transformers

hf-experiments-128

On πŸ”₯ πŸ”¬ Experiments πŸ†•

If you are interested in Text Generation, we have just added GPT-J 6B that has a PPL of 3.99 and ACC of 69.7%. We also provide *GPT-Neo 1.3B, 2.7B as well as smaller 350M and 125M parameters. Check here for evaluations.

πŸ€— Huggingface πŸ”¬ Experiments

The following experiments available through HF models are supported:

  • GPT-J 6B: GPT-J 6B is a transformer model trained using Ben Wang's Mesh Transformer JAX. πŸ†• πŸ”₯
  • HuBERT: Self-supervised representation learning for speech recognition, generation, and compression
  • zeroshot - NLI-based Zero Shot Text Classification (ZSL)
  • nrot - Numerical reasoning over text (NRoT) pretrained models (NT5)
  • vit - Vision Transformer (ViT) model pre-trained on ImageNet
  • bigbird - Google sparse-attention based transformer which extends Transformer based models to much longer sequences
  • msmarco - Sentence BERT's MSMarco for Semantic Search and Retrieve & Re-Rank πŸ”₯
  • luke - LUKE is a RoBERTa model that does named entity recognition, extractive and cloze-style question answering, entity typing, and relation classification πŸ”₯
  • colbert - Model is based on ColBERT: Efficient and Effective Passage Search via Contextualized Late Interaction over BERT
  • audioseg - Pyannote audio segmentation and speaker diarization πŸ”₯
  • asr - automatic speech recognition
  • gpt_neo - EleutherAI's replication of the GPT-3 πŸ”₯
  • bert BERT Transformer: Masked Language Modeling, Next Sentence Prediction, Extractive Question Answering πŸ”₯
  • summarization - text summarization
  • translation - text multiple languages translation
  • sentiment - sentiment analysis
  • emotions - emotions detection
  • pokemon - PokΓ©mon 🐣 🐒 πŸ¦€ πŸ„ πŸ¦‡πŸ¦‚ generator based on russian RuDALL-E πŸ†• πŸ”₯

Not-Huggingface πŸ”¬ Experiments

We propose some additional experiments currently not avaiable on HF models' hub

  • audioset - YamNet Image classification and VGGish Image embedding on AudioSet Youtube Corpus
  • genre - Generative ENtity REtrieval πŸ”₯
  • mlpvision - MLP Mixex, ResMLP, Perceiver models for Computer Vision
  • fewnerd - Few-NERD: Not Only a Few-shot NER Dataset πŸ”₯
  • skweak - Weak supervision for NLP πŸ”₯
  • projected_gan - NeurIPS 2021 "Projected GANs Converge Faster"
  • fasttext - FastText a library for efficient learning of word representations and sentence classification.
  • whisper, general-purpose speech recognition, multilingual speech recognition, speech translation, spoken language identification, and voice activity detection model πŸ†• πŸ”₯
  • alphatensor Discovering faster matrix multiplication algorithms with reinforcement learning. Nature 610 (2022) πŸ†• πŸ”₯

How to build

To build experiments run

./build.sh

To build experiments with GPU run

./build.sh gpu

How to run

To run an experiment run

./run.sh [experiment_name] [gpu|cpu] [cache_dir_folder]

To run an experiment on GPU run

./run.sh [experiment_name] gpu [cache_dir_folder]

The experiment_name field is among the following supported experiment names, while the cache_dir_folder parameter is the directorty where to cache models files. See later about this.

How to debug

To debug the code, without running any experiment

./debug.sh
root@d2f0e8a5ec76:/app# 

To debug for GPU run

./debug.sh gpu

This will enter the running image hfexperiments. You can now run python scripts manually, like

root@d2f0e8a5ec76:/app# python src/asr/run.py

NOTE. For preconfigured experiments, please run the run.py script from the main folder /app, as the cache directories are following that path, so like python src/asr/run.py

Dependencies

We are up-to-date with the latest transformers, Pytorch, tensorflow and Keras models, and we also provide most common ML libraries:

Package                 Version     
----------------------- ------------
transformers            4.5.1
tokenizers              0.10.2 
torch                   1.8.1
tensorflow              2.4.1
Keras                   2.4.3
pytorch-lightning       1.2.10
numpy                   1.19.5
tensorboard             2.4.1
sentencepiece           0.1.95
pyannote.core           4.1
librosa                 0.8.0
matplotlib              3.4.1
pandas                  1.2.4 
scikit-learn            0.24.2
scipy                   1.6.3 

Common Dependencies are defined in the requirements.txt file and currently are

torch
tensorflow
keras
transformers
sentencepiece
soundfile

Dev dependencies

Due to high rate of πŸ†• models pushed to the Huggingface models hub, we provide a requirements-dev.txt in order to install the latest master branch of transformers:

./debug.sh
pip install -r requirements-dev.txt

Experiment Dependencies

Experiment level dependencies are specified in app folder requirements.txt file like src/asr/requirements.txt for asr experiment.

Models files

Where are models files saved? Models files are typically big. It's preferable to save them to a custom folder like an external HDD of a shared disk. For this reason a docker environment variable cache_dir can specified at run:

./run.sh emotions models/

the models folder will be assigned to the cache_dir variable to be used as default alternative location to download pretrained models. A os.getenv("cache_dir") will be used to retrieve the environemnt variable in the code.

Additional models files

Some experiments require additional models to be downloaed, not currently available through Huggingface model's hub, therefore a courtesy download script has been provided in the experiment's folder like, genre/models.sh for the following experiments:

  • audioset
  • genre
  • megatron

We do not automatically download these files, so please run in debug mode with debug.sh and download the models manually, before running those experiments. The download shall be done once, and the models files will be placed in the models' cache folder specified by environment variable cache_dir as it happens for the Huggingface's Model Hub.

Contributors

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].