All Projects → dhgrs → chainer-ClariNet

dhgrs / chainer-ClariNet

Licence: other
A Chainer implementation of ClariNet.

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to chainer-ClariNet

Wavenet
WaveNet implementation with chainer
Stars: ✭ 53 (+17.78%)
Mutual labels:  chainer, wavenet
chainer-Fast-WaveNet
A Chainer implementation of Fast WaveNet(mel-spectrogram vocoder).
Stars: ✭ 33 (-26.67%)
Mutual labels:  chainer, wavenet
QPPWG
Quasi-Periodic Parallel WaveGAN Pytorch implementation
Stars: ✭ 41 (-8.89%)
Mutual labels:  wavenet, parallel-wavenet-vocoder
Chainer Vq Vae
A Chainer implementation of VQ-VAE.
Stars: ✭ 77 (+71.11%)
Mutual labels:  chainer, wavenet
wavenet
Audio source separation (mixture to vocal) using the Wavenet
Stars: ✭ 20 (-55.56%)
Mutual labels:  wavenet
chainer-pix2pix
Chainer implementation for Image-to-Image Translation Using Conditional Adversarial Networks
Stars: ✭ 40 (-11.11%)
Mutual labels:  chainer
Visual-Attention-Model
Chainer implementation of Deepmind's Visual Attention Model paper
Stars: ✭ 27 (-40%)
Mutual labels:  chainer
ChainerPruner
ChainerPruner: Channel Pruning framework for Chainer
Stars: ✭ 21 (-53.33%)
Mutual labels:  chainer
tutorials
Introduction to Deep Learning: Chainer Tutorials
Stars: ✭ 68 (+51.11%)
Mutual labels:  chainer
superresolution gan
Chainer implementation of Photo-Realistic Single Image Super-Resolution Using a Generative Adversarial Network
Stars: ✭ 50 (+11.11%)
Mutual labels:  chainer
rocgan
Chainer implementation of the paper Robust Conditional Generative Adversarial Networks
Stars: ✭ 15 (-66.67%)
Mutual labels:  chainer
BMI219-2017-ProteinFolding
UCSF BMI219 Deep Learning (2017), Coding example (Prediction of protein folding with RNN and CNN)
Stars: ✭ 14 (-68.89%)
Mutual labels:  chainer
chainer-graph-cnn
Chainer implementation of 'Convolutional Neural Networks on Graphs with Fast Localized Spectral Filtering' (https://arxiv.org/abs/1606.09375)
Stars: ✭ 67 (+48.89%)
Mutual labels:  chainer
chainer2pytorch
Converts Chainer modules to PyTorch, parameters included.
Stars: ✭ 36 (-20%)
Mutual labels:  chainer
constant-memory-waveglow
PyTorch implementation of NVIDIA WaveGlow with constant memory cost.
Stars: ✭ 36 (-20%)
Mutual labels:  wavenet
birdsong-generation-project
Generating birdsong with WaveNet
Stars: ✭ 26 (-42.22%)
Mutual labels:  wavenet
chainer-param-monitor
Monitor parameter and gradient statistics during neural network training with Chainer
Stars: ✭ 13 (-71.11%)
Mutual labels:  chainer
chainer-fcis
[This project has moved to ChainerCV] Chainer Implementation of Fully Convolutional Instance-aware Semantic Segmentation
Stars: ✭ 45 (+0%)
Mutual labels:  chainer
convolutional seq2seq
fairseq: Convolutional Sequence to Sequence Learning (Gehring et al. 2017) by Chainer
Stars: ✭ 63 (+40%)
Mutual labels:  chainer
chainer-sort
Simple, Online, Realtime Tracking of Multiple Objects (SORT) implementation for Chainer and ChainerCV.
Stars: ✭ 20 (-55.56%)
Mutual labels:  chainer

chainer-ClariNet

A Chainer implementation of ClariNet( https://arxiv.org/abs/1807.07281 ).

Results

Autoregressive WaveNet(Single Gaussian ver.) trained with VCTK Corpus

https://nana-music.com/sounds/04027269/

Student Gaussian IAF trained with LJ-Speech

https://nana-music.com/sounds/043ba7b4/

Requirements

I trained and generated with

  • python(3.5.2)
  • chainer (5.0.0b4)
  • librosa (0.6.2)
  • matplotlib (2.2.3)
  • tqdm (4.25.0)

Usage

download dataset

You can download VCTK Corpus(en multi speaker)/LJ-Speech(en single speaker) very easily via my repository.

set parameters

Almost parameters in params.py and teacher_params.py are same as params.py in my other repositories like VQ-VAE. If you modified params.py in AutoregressiveWavenet, you have to replace teacher_params.py with it to train student.

training

You can use same command in each directory.

(without GPU)
python train.py

(with GPU #n)
python train.py -g n

You can resume snapshot and restart training like below.(Now support AutoregressiveWaveNet only)

python train.py -r snapshot_iter_100000

Other arguments -f and -p are parameters for multiprocess in preprocessing. -f means the number of prefetch and -p means the number of processes. I highly recommend to modify -f to large number like 64. If GPU-Util is stil low, modify -p to large number like 8.

generating

python generate.py -i <input file> -o <output file> -m <trained model>

If you don't set -o, default file name result.wav is used. If you don't set -s, the speaker is same as input file that got from filepath.

Caution

I only check the results for

  • Autoregressive WaveNet(Single Gaussian ver.)
  • Student Gaussian IAF
Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].