All Projects → mravanelli → Pyspeechrev

mravanelli / Pyspeechrev

This python code performs an efficient speech reverberation starting from a dataset of close-talking speech signals and a collection of acoustic impulse responses.

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to Pyspeechrev

Iflytek awaken asr
use iflytek's technology to realize awaken and order recognition
Stars: ✭ 53 (-28.38%)
Mutual labels:  speech-recognition
Papers
A list of paper, books and sites for various different topics related to machine learning and deep learning along with various field in which it is implemented
Stars: ✭ 63 (-14.86%)
Mutual labels:  speech-recognition
Coordconv
Pytorch implementation of "An intriguing failing of convolutional neural networks and the CoordConv solution" - https://arxiv.org/abs/1807.03247
Stars: ✭ 72 (-2.7%)
Mutual labels:  convolution
Biglittlenet
Official repository for Big-Little Net
Stars: ✭ 57 (-22.97%)
Mutual labels:  speech-recognition
Angle
⦠ Angle: new speakable syntax for python 💡
Stars: ✭ 61 (-17.57%)
Mutual labels:  speech-recognition
Speech ai
Simple speech linguistic AI with Python
Stars: ✭ 66 (-10.81%)
Mutual labels:  speech-recognition
Parrots
Automatic Speech Recognition(ASR), Text-To-Speech(TTS) engine for Chinese.
Stars: ✭ 48 (-35.14%)
Mutual labels:  speech-recognition
Unityandroidspeechrecognition
This repository is a Unity plugin for Android Speech Recognition (based on Java implementation)
Stars: ✭ 73 (-1.35%)
Mutual labels:  speech-recognition
Dragonfire
the open-source virtual assistant for Ubuntu based Linux distributions
Stars: ✭ 1,120 (+1413.51%)
Mutual labels:  speech-recognition
Asr benchmark
Program to benchmark various speech recognition APIs
Stars: ✭ 71 (-4.05%)
Mutual labels:  speech-recognition
Dolphinattack
Inaudible Voice Commands
Stars: ✭ 57 (-22.97%)
Mutual labels:  speech-recognition
Audio Pretrained Model
A collection of Audio and Speech pre-trained models.
Stars: ✭ 61 (-17.57%)
Mutual labels:  speech-recognition
Openasr
A pytorch based end2end speech recognition system.
Stars: ✭ 69 (-6.76%)
Mutual labels:  speech-recognition
Active Convolution
Active Convolution
Stars: ✭ 56 (-24.32%)
Mutual labels:  convolution
Android Speech Recognition
Continuous speech recognition library for Android with options to use GoogleVoiceIme dialog and offline mode.
Stars: ✭ 72 (-2.7%)
Mutual labels:  speech-recognition
Anomaly detection
This is a times series anomaly detection algorithm, implemented in Python, for catching multiple anomalies. It uses a moving average with an extreme student deviate (ESD) test to detect anomalous points.
Stars: ✭ 50 (-32.43%)
Mutual labels:  convolution
Fft Conv Pytorch
Implementation of 1D, 2D, and 3D FFT convolutions in PyTorch. Much faster than direct convolutions for large kernel sizes.
Stars: ✭ 65 (-12.16%)
Mutual labels:  convolution
Scidart
Multiplatform scientific computing for Dart
Stars: ✭ 73 (-1.35%)
Mutual labels:  convolution
Nativescript Speech Recognition
💬 Speech to text, using the awesome engines readily available on the device.
Stars: ✭ 72 (-2.7%)
Mutual labels:  speech-recognition
Patter
speech-to-text in pytorch
Stars: ✭ 71 (-4.05%)
Mutual labels:  speech-recognition

pySpeechRev

This python code performs an efficient speech reverberation starting from a dataset of close-talking speech signals and a collection of acoustic impulse responses.

The reverberated signal y[n] is computed in the following way:

y[n]=x[n] * h[n]

where x[n] is the clean signal and * is the convolutional operator.

The script takes in input the following arguments:

  • in_folder: folder where the original close-talk dataset is stored.
  • out_folder: folder where the reverberated dataset will be stored.
  • list.txt : it is a text file where each row should contain: original_wav_file IR_file.

Before run it, make sure you have all the needed python packages. In particular:

  • pysoundfile: pip install pysoundfile
  • numpy
  • scipy

Example:

python pySpeechRev.py clean_examples/ rev_examples/ list.txt

Note that to have meaningful and realistic results, both the impulse responses and the clean speech signal must be sampled at the same sampling rate (e.g., 16 kHz - 16 kHz).

Reverberated TIMIT

To create a reverberated version of TIMIT do the following steps:

  • Make sure you have the TIMIT dataset. If not, it can be downloaded from the LDC website (https://catalog.ldc.upenn.edu/LDC93S1).
  • Change lst_TIMIT.txt according to the paths of your TIMIT Dataset
  • Run:
python pySpeechRev.py $path_TIMIT  $path_TIMIT_rev lst_TIMIT.txt

The current version of TIMIT has been contaminated with some high-quality impulse responses of the DIRHA-English Dataset [3].

Tested on: Python 2.7, Ubuntu

This code has been used in the following papers (please cite them if you use this code):

[1] M. Ravanelli, P. Svaizer, M. Omologo, "Realistic Multi-Microphone Data Simulation for Distant Speech Recognition", in Proceedings of Interspeech 2016. https://arxiv.org/abs/1711.09470

[2] M. Ravanelli, M. Omologo, "Contaminated speech training methods for robust DNN-HMM distant speech recognition", in Proceedings of INTERSPEECH 2015. https://arxiv.org/abs/1710.03538

[3] M. Ravanelli, M. Omologo, "The DIRHA-English corpus and related tasks for distant-speech recognition in domestic environments", in Proceedings of ASRU 2015. https://arxiv.org/abs/1710.02560

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].