All Projects → NVIDIA → Flowtron

NVIDIA / Flowtron

Licence: apache-2.0
Flowtron is an auto-regressive flow-based generative network for text to speech synthesis with control over speech variation and style transfer

Projects that are alternatives of or similar to Flowtron

Wavegrad
Implementation of Google Brain's WaveGrad high-fidelity vocoder (paper: https://arxiv.org/pdf/2009.00713.pdf). First implementation on GitHub.
Stars: ✭ 245 (-55.13%)
Mutual labels:  jupyter-notebook, speech-synthesis
Cs224n Gpu That Talks
Attention, I'm Trying to Speak: End-to-end speech synthesis (CS224n '18)
Stars: ✭ 52 (-90.48%)
Mutual labels:  jupyter-notebook, speech-synthesis
Parallelwavegan
Unofficial Parallel WaveGAN (+ MelGAN & Multi-band MelGAN) with Pytorch
Stars: ✭ 682 (+24.91%)
Mutual labels:  jupyter-notebook, speech-synthesis
Waveflow
A PyTorch implementation of "WaveFlow: A Compact Flow-based Model for Raw Audio"
Stars: ✭ 95 (-82.6%)
Mutual labels:  jupyter-notebook, speech-synthesis
Nemo
NeMo: a toolkit for conversational AI
Stars: ✭ 3,685 (+574.91%)
Mutual labels:  jupyter-notebook, speech-synthesis
Pytorch Dc Tts
Text to Speech with PyTorch (English and Mongolian)
Stars: ✭ 122 (-77.66%)
Mutual labels:  jupyter-notebook, speech-synthesis
Tf Wavenet vocoder
Wavenet and its applications with Tensorflow
Stars: ✭ 58 (-89.38%)
Mutual labels:  jupyter-notebook, speech-synthesis
Tacotron pytorch
PyTorch implementation of Tacotron speech synthesis model.
Stars: ✭ 242 (-55.68%)
Mutual labels:  jupyter-notebook, speech-synthesis
Gantts
PyTorch implementation of GAN-based text-to-speech synthesis and voice conversion (VC)
Stars: ✭ 460 (-15.75%)
Mutual labels:  jupyter-notebook, speech-synthesis
Photomosaic
Creating fun photomosaics, GIFs, and murals from your family pictures using ML & similarity search
Stars: ✭ 540 (-1.1%)
Mutual labels:  jupyter-notebook
Tensorrt
TensorFlow/TensorRT integration
Stars: ✭ 543 (-0.55%)
Mutual labels:  jupyter-notebook
Keras Openface
Keras-OpenFace is a project converting OpenFace from Torch implementation to a Keras version
Stars: ✭ 538 (-1.47%)
Mutual labels:  jupyter-notebook
Pose Hg Train
Training and experimentation code used for "Stacked Hourglass Networks for Human Pose Estimation"
Stars: ✭ 541 (-0.92%)
Mutual labels:  jupyter-notebook
Chinese models for spacy
SpaCy 中文模型 | Models for SpaCy that support Chinese
Stars: ✭ 543 (-0.55%)
Mutual labels:  jupyter-notebook
Attention Networks For Classification
Hierarchical Attention Networks for Document Classification in PyTorch
Stars: ✭ 540 (-1.1%)
Mutual labels:  jupyter-notebook
Hate Speech And Offensive Language
Repository for the paper "Automated Hate Speech Detection and the Problem of Offensive Language", ICWSM 2017
Stars: ✭ 543 (-0.55%)
Mutual labels:  jupyter-notebook
Notebooks
Some sample IPython notebooks for scikit-learn
Stars: ✭ 539 (-1.28%)
Mutual labels:  jupyter-notebook
Python intro
Jupyter notebooks in Russian. Introduction to Python, basic algorithms and data structures
Stars: ✭ 538 (-1.47%)
Mutual labels:  jupyter-notebook
Sentiment analysis fine grain
Multi-label Classification with BERT; Fine Grained Sentiment Analysis from AI challenger
Stars: ✭ 546 (+0%)
Mutual labels:  jupyter-notebook
Pierian Data Complete Python 3 Bootcamp
Stars: ✭ 544 (-0.37%)
Mutual labels:  jupyter-notebook

Flowtron

Flowtron: an Autoregressive Flow-based Network for Text-to-Mel-spectrogram Synthesis

Rafael Valle, Kevin Shih, Ryan Prenger and Bryan Catanzaro

In our recent paper we propose Flowtron: an autoregressive flow-based generative network for text-to-speech synthesis with control over speech variation and style transfer. Flowtron borrows insights from Autoregressive Flows and revamps Tacotron in order to provide high-quality and expressive mel-spectrogram synthesis. Flowtron is optimized by maximizing the likelihood of the training data, which makes training simple and stable. Flowtron learns an invertible mapping of data to a latent space that can be manipulated to control many aspects of speech synthesis (pitch, tone, speech rate, cadence, accent).

Our mean opinion scores (MOS) show that Flowtron matches state-of-the-art TTS models in terms of speech quality. In addition, we provide results on control of speech variation, interpolation between samples and style transfer between speakers seen and unseen during training.

Visit our website for audio samples.

Pre-requisites

  1. NVIDIA GPU + CUDA cuDNN

Setup

  1. Clone this repo: git clone https://github.com/NVIDIA/flowtron.git
  2. CD into this repo: cd flowtron
  3. Initialize submodule: git submodule update --init; cd tacotron2; git submodule update --init
  4. Install PyTorch
  5. Install python requirements or build docker image
    • Install python requirements: pip install -r requirements.txt

Training from scratch

  1. Update the filelists inside the filelists folder to point to your data
  2. Train using the attention prior until attention looks good python train.py -c config.json -p train_config.output_directory=outdir data_config.use_attn_prior=1
  3. Resume training without the attention prior python train.py -c config.json -p train_config.output_directory=outdir data_config.use_attn_prior=0 train_config.checkpoint_path=model_niters
  4. (OPTIONAL) If the gate layer is overfitting once done training, train just the gate layer from scratch python train.py -c config.json -p train_config.output_directory=outdir train_config.checkpoint_path=model_niters data_config.use_attn_prior=0 train_config.ignore_layers='["flows.1.ar_step.gate_layer.linear_layer.weight","flows.1.ar_step.gate_layer.linear_layer.bias"]' train_config.finetune_layers='["flows.1.ar_step.gate_layer.linear_layer.weight","flows.1.ar_step.gate_layer.linear_layer.bias"]'
  5. (OPTIONAL) tensorboard --logdir=outdir/logdir

Training using a pre-trained model

Training using a pre-trained model can lead to faster convergence. Dataset dependent layers can be ignored

  1. Download our published Flowtron LJS, Flowtron LibriTTS or Flowtron LibriTTS2K model
  2. python train.py -c config.json -p train_config.ignore_layers=["speaker_embedding.weight"] train_config.checkpoint_path="models/flowtron_ljs.pt"

Fine-tuning for few-shot speech synthesis

  1. Download our published Flowtron LibriTTS2K model
  2. python train.py -c config.json -p train_config.finetune_layers=["speaker_embedding.weight"] train_config.checkpoint_path="models/flowtron_libritts2k.pt"

Multi-GPU (distributed) and Automatic Mixed Precision Training (AMP)

  1. python -m torch.distributed.launch --use_env --nproc_per_node=NUM_GPUS_YOU_HAVE train.py -c config.json -p train_config.output_directory=outdir train_config.fp16=true

Inference demo

  1. python inference.py -c config.json -f models/flowtron_ljs.pt -w models/waveglow_256channels_v4.pt -t "It is well know that deep generative models have a deep latent space!" -i 0

Related repos

WaveGlow Faster than real time Flow-based Generative Network for Speech Synthesis

Acknowledgements

This implementation uses code from the following repos: Keith Ito, Prem Seetharaman and Liyuan Liu as described in our code.

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].