All Projects → dhgrs → Pytorch Uniwavenet

dhgrs / Pytorch Uniwavenet

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to Pytorch Uniwavenet

Vq Vae Speech
PyTorch implementation of VQ-VAE + WaveNet by [Chorowski et al., 2019] and VQ-VAE on speech signals by [van den Oord et al., 2017]
Stars: ✭ 187 (+523.33%)
Mutual labels:  wavenet
chainer-Fast-WaveNet
A Chainer implementation of Fast WaveNet(mel-spectrogram vocoder).
Stars: ✭ 33 (+10%)
Mutual labels:  wavenet
Pycadl
Python package with source code from the course "Creative Applications of Deep Learning w/ TensorFlow"
Stars: ✭ 356 (+1086.67%)
Mutual labels:  wavenet
birdsong-generation-project
Generating birdsong with WaveNet
Stars: ✭ 26 (-13.33%)
Mutual labels:  wavenet
constant-memory-waveglow
PyTorch implementation of NVIDIA WaveGlow with constant memory cost.
Stars: ✭ 36 (+20%)
Mutual labels:  wavenet
hifigan-denoiser
HiFi-GAN: High Fidelity Denoising and Dereverberation Based on Speech Deep Features in Adversarial Networks
Stars: ✭ 88 (+193.33%)
Mutual labels:  wavenet
Source Separation Wavenet
A neural network for end-to-end music source separation
Stars: ✭ 185 (+516.67%)
Mutual labels:  wavenet
Parallelwavegan
Unofficial Parallel WaveGAN (+ MelGAN & Multi-band MelGAN) with Pytorch
Stars: ✭ 682 (+2173.33%)
Mutual labels:  wavenet
chainer-ClariNet
A Chainer implementation of ClariNet.
Stars: ✭ 45 (+50%)
Mutual labels:  wavenet
Time Series Prediction
A collection of time series prediction methods: rnn, seq2seq, cnn, wavenet, transformer, unet, n-beats, gan, kalman-filter
Stars: ✭ 351 (+1070%)
Mutual labels:  wavenet
wavenet
Audio source separation (mixture to vocal) using the Wavenet
Stars: ✭ 20 (-33.33%)
Mutual labels:  wavenet
QPPWG
Quasi-Periodic Parallel WaveGAN Pytorch implementation
Stars: ✭ 41 (+36.67%)
Mutual labels:  wavenet
Pytorchwavenetvocoder
WaveNet-Vocoder implementation with pytorch.
Stars: ✭ 269 (+796.67%)
Mutual labels:  wavenet
wavenet-like-vocoder
Basic wavenet and fftnet vocoder model.
Stars: ✭ 20 (-33.33%)
Mutual labels:  wavenet
Flowavenet
A Pytorch implementation of "FloWaveNet: A Generative Flow for Raw Audio"
Stars: ✭ 471 (+1470%)
Mutual labels:  wavenet
Seriesnet
Time series prediction using dilated causal convolutional neural nets (temporal CNN)
Stars: ✭ 185 (+516.67%)
Mutual labels:  wavenet
ttslearn
ttslearn: Library for Pythonで学ぶ音声合成 (Text-to-speech with Python)
Stars: ✭ 158 (+426.67%)
Mutual labels:  wavenet
Wavenet Stt
An end-to-end speech recognition system with Wavenet. Built using C++ and python.
Stars: ✭ 18 (-40%)
Mutual labels:  wavenet
Speech Denoising Wavenet
A neural network for end-to-end speech denoising
Stars: ✭ 516 (+1620%)
Mutual labels:  wavenet
Clarinet
A Pytorch Implementation of ClariNet
Stars: ✭ 273 (+810%)
Mutual labels:  wavenet

pytorch-UniWaveNet

A PyTorch implementetion of Uni-WaveNet( https://zhuanlan.zhihu.com/p/44702855 ). Uni-WaveNet is derived from Parallel WaveNet, it can be trained without Teacher WaveNet.

Generated samples

I uploaded generated samples to SoundCloud. https://soundcloud.com/dhgrs/sets/uni-wavenet

And uploaded pretrained model for LJSpeech to Google Drive, same setting as NVIDIA's Tacotron2 implementaiton. https://drive.google.com/drive/folders/1BqzltOT9u3358nQgPolRA511J7cGJGlI?usp=sharing

Requirements

  • Python3
  • PyTorch(>0.4.1)
  • tensorboardX
  • numpy
  • libsosa
  • tqdm

CAUTION: The interface of PyTorch's STFT API has changed at ver. 0.4.1. So you have to use 0.4.1 or later.

Usage

  1. Download dataset

This implementation can train with LJSpeech, English single speaker corpus or VCTK-Corpus, English multi speaker corpus. And you can download them very easily via my repository.

  1. Set parameters

Hyperparameters are in params.py. You have to change root to the directory you download the dataset. If you can not understand some parameters, please open an issue.

  1. Run training
# without GPU
python3 train.py

# with GPU
python3 train.py --use_cuda

If you want to restart training with snapshot, use options like below.

python3 train.py --use_cuda -e path/to/encoder.pth -w path/to/wavenet.pth -o path/to/optimizer.pth -i iteration_to_start_at

You can use TensorBoard to visualize training. Also can listen the generated samples during training.

  1. Generate
python3 generate.py -e path/to/encoder.pth -w path/to/wavenet.pth -i path/to/input.wav -l length_to_generate(sec)

CAUTION: There is a bug on PyTorch and my implementation. Now you can generate samples on ONLY GPU. You will get an error if you try to generate on CPU. Related issue which I opened is here.

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].