All Projects → k2kobayashi → Sprocket

k2kobayashi / Sprocket

Licence: mit
Voice Conversion Tool Kit

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to Sprocket

editts
Official implementation of EdiTTS: Score-based Editing for Controllable Text-to-Speech
Stars: ✭ 74 (-82.59%)
Mutual labels:  speech-synthesis
Parakeet
PAddle PARAllel text-to-speech toolKIT (supporting WaveFlow, WaveNet, Transformer TTS and Tacotron2)
Stars: ✭ 279 (-34.35%)
Mutual labels:  speech-synthesis
Multilingual text to speech
An implementation of Tacotron 2 that supports multilingual experiments with parameter-sharing, code-switching, and voice cloning.
Stars: ✭ 324 (-23.76%)
Mutual labels:  speech-synthesis
Tacotron pytorch
Tacotron implementation of pytorch
Stars: ✭ 12 (-97.18%)
Mutual labels:  speech-synthesis
Comprehensive-Tacotron2
PyTorch Implementation of Google's Natural TTS Synthesis by Conditioning WaveNet on Mel Spectrogram Predictions. This implementation supports both single-, multi-speaker TTS and several techniques to enforce the robustness and efficiency of the model.
Stars: ✭ 22 (-94.82%)
Mutual labels:  speech-synthesis
Pysptk
A python wrapper for Speech Signal Processing Toolkit (SPTK).
Stars: ✭ 297 (-30.12%)
Mutual labels:  speech-synthesis
Fre-GAN-pytorch
Fre-GAN: Adversarial Frequency-consistent Audio Synthesis
Stars: ✭ 73 (-82.82%)
Mutual labels:  speech-synthesis
Voice Builder
An opensource text-to-speech (TTS) voice building tool
Stars: ✭ 362 (-14.82%)
Mutual labels:  speech-synthesis
Pytorchwavenetvocoder
WaveNet-Vocoder implementation with pytorch.
Stars: ✭ 269 (-36.71%)
Mutual labels:  speech-synthesis
Hifi Gan
HiFi-GAN: Generative Adversarial Networks for Efficient and High Fidelity Speech Synthesis
Stars: ✭ 325 (-23.53%)
Mutual labels:  speech-synthesis
EmotionalConversionStarGAN
This repository contains code to replicate results from the ICASSP 2020 paper "StarGAN for Emotional Speech Conversion: Validated by Data Augmentation of End-to-End Emotion Recognition".
Stars: ✭ 92 (-78.35%)
Mutual labels:  speech-synthesis
voice-conversion
an tutorial implement of voice conversion using pytorch
Stars: ✭ 26 (-93.88%)
Mutual labels:  speech-synthesis
Nnmnkwii
Library to build speech synthesis systems designed for easy and fast prototyping.
Stars: ✭ 308 (-27.53%)
Mutual labels:  speech-synthesis
voicekit-examples
Examples on how to use Tinkoff Voicekit
Stars: ✭ 35 (-91.76%)
Mutual labels:  speech-synthesis
Espeak
eSpeak NG is an open source speech synthesizer that supports 101 languages and accents.
Stars: ✭ 339 (-20.24%)
Mutual labels:  speech-synthesis
leon
🧠 Leon is your open-source personal assistant.
Stars: ✭ 8,560 (+1914.12%)
Mutual labels:  speech-synthesis
Glow Tts
A Generative Flow for Text-to-Speech via Monotonic Alignment Search
Stars: ✭ 284 (-33.18%)
Mutual labels:  speech-synthesis
Espnet
End-to-End Speech Processing Toolkit
Stars: ✭ 4,533 (+966.59%)
Mutual labels:  speech-synthesis
Libfaceid
libfaceid is a research framework for prototyping of face recognition solutions. It seamlessly integrates multiple detection, recognition and liveness models w/ speech synthesis and speech recognition.
Stars: ✭ 354 (-16.71%)
Mutual labels:  speech-synthesis
Cognitive Speech Tts
Microsoft Text-to-Speech API sample code in several languages, part of Cognitive Services.
Stars: ✭ 312 (-26.59%)
Mutual labels:  speech-synthesis

Python Version Build Status Coverage Status PyPI version MIT License

sprocket

Voice conversion software - Voice conversion (VC) is a technique to convert a speaker identity of a source speaker into that of a target speaker. This software enables the users to develop a traditional VC system based on a Gaussian mixture model (GMM) and a vocoder-free VC system based on a differential GMM (DIFFGMM) using a parallel dataset of the source and target speakers.

Paper and slide

  • K. Kobayashi, T. Toda, "sprocket: Open-Source Voice Conversion Software," Proc. Odyssey, pp. 203-210, June 2018. [paper]

  • T. Toda, "Hands on Voice Conversion," Speech Processing Courses in Crete (SPCC), July 2018. [slide]

Conversion samples

Purpose

Reproduce the typical VC systems

This software was developed to make it possible for the users to easily build the VC systems by only preparing a parallel dataset of the desired source and target speakers and executing example scripts. The following VC methods were implemented as the typical VC methods.

Traditional VC method based on GMM

  • T. Toda, A.W. Black, K. Tokuda, "Voice conversion based on maximum likelihood estimation of spectral parameter trajectory," IEEE Transactions on Audio, Speech and Language Processing, Vol. 15, No. 8, pp. 2222-2235, Nov. 2007.

Vocoder-free VC method based on DIFFGMM

  • K. Kobayashi, T. Toda, S. Nakamura, "F0 transformation techniques for statistical voice conversion with direct waveform modification with spectral differential," Proc. IEEE SLT, pp. 693-700, Dec. 2016.

Supply Python3 VC library

To make it possible to easily develop VC-based applications using Python (Python3), the VC library is also supplied, including several interfaces, such as acoustic feature analysis/synthesis, acoustic feature modeling, acoustic feature conversion, and waveform modification. For the details of the VC library, please see sprocket documents in (coming soon).

Installation & Run

Please use Python3.

Current stable version

Ver. 0.18.4

Install sprocket

pip install numpy==1.15.4 cython  # for dependency
pip install sprocket-vc

Run example

See VC example

REPORTING BUGS

For any questions or issues please visit:

https://github.com/k2kobayashi/sprocket/issues

COPYRIGHT

Copyright (c) 2020 Kazuhiro KOBAYASHI

Released under the MIT license

https://opensource.org/licenses/mit-license.php

ACKNOWLEDGEMENTS

Thank you @r9y9 and @tats-u for lots of contributions and encouragement helps before release.

Who we are

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].