All Projects → arjo129 → Uspeech

arjo129 / Uspeech

Licence: mit
Speech recognition toolkit for the arduino

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to Uspeech

Pulsesensorstarterproject
The Best Way to Get Started with your PulseSensor and Arduino
Stars: ✭ 38 (-91.52%)
Mutual labels:  arduino, signal
awesome-keyword-spotting
This repository is a curated list of awesome Speech Keyword Spotting (Wake-Up Word Detection).
Stars: ✭ 150 (-66.52%)
Mutual labels:  speech-recognition, speech-processing
UHV-OTS-Speech
A data annotation pipeline to generate high-quality, large-scale speech datasets with machine pre-labeling and fully manual auditing.
Stars: ✭ 94 (-79.02%)
Mutual labels:  speech-recognition, speech-processing
Zzz Retired openstt
RETIRED - OpenSTT is now retired. If you would like more information on Mycroft AI's open source STT projects, please visit:
Stars: ✭ 146 (-67.41%)
Mutual labels:  speech-recognition, speech-processing
Speech-Backbones
This is the main repository of open-sourced speech technology by Huawei Noah's Ark Lab.
Stars: ✭ 205 (-54.24%)
Mutual labels:  speech-recognition, speech-processing
Speechbrain.github.io
The SpeechBrain project aims to build a novel speech toolkit fully based on PyTorch. With SpeechBrain users can easily create speech processing systems, ranging from speech recognition (both HMM/DNN and end-to-end), speaker recognition, speech enhancement, speech separation, multi-microphone speech processing, and many others.
Stars: ✭ 242 (-45.98%)
Mutual labels:  speech-recognition, speech-processing
react-native-spokestack
Spokestack: give your React Native app a voice interface!
Stars: ✭ 53 (-88.17%)
Mutual labels:  speech-recognition, speech-processing
Pncc
A implementation of Power Normalized Cepstral Coefficients: PNCC
Stars: ✭ 40 (-91.07%)
Mutual labels:  speech-recognition, speech-processing
open-speech-corpora
💎 A list of accessible speech corpora for ASR, TTS, and other Speech Technologies
Stars: ✭ 841 (+87.72%)
Mutual labels:  speech-recognition, speech-processing
QuantumSpeech-QCNN
IEEE ICASSP 21 - Quantum Convolution Neural Networks for Speech Processing and Automatic Speech Recognition
Stars: ✭ 71 (-84.15%)
Mutual labels:  speech-recognition, speech-processing
Nonautoreggenprogress
Tracking the progress in non-autoregressive generation (translation, transcription, etc.)
Stars: ✭ 118 (-73.66%)
Mutual labels:  speech-recognition, speech-processing
UniSpeech
UniSpeech - Large Scale Self-Supervised Learning for Speech
Stars: ✭ 224 (-50%)
Mutual labels:  speech-recognition, speech-processing
Keras Sincnet
Keras (tensorflow) implementation of SincNet (Mirco Ravanelli, Yoshua Bengio - https://github.com/mravanelli/SincNet)
Stars: ✭ 47 (-89.51%)
Mutual labels:  speech-recognition, speech-processing
Sinewave
Sine wave library for Arduino. Produces a sine signal and passes the output to a PWM pin.
Stars: ✭ 11 (-97.54%)
Mutual labels:  arduino, signal
Formant Analyzer
iOS application for finding formants in spoken sounds
Stars: ✭ 43 (-90.4%)
Mutual labels:  speech-recognition, speech-processing
torchsubband
Pytorch implementation of subband decomposition
Stars: ✭ 63 (-85.94%)
Mutual labels:  speech-recognition, speech-processing
Awesome Diarization
A curated list of awesome Speaker Diarization papers, libraries, datasets, and other resources.
Stars: ✭ 673 (+50.22%)
Mutual labels:  speech-recognition, speech-processing
Sincnet
SincNet is a neural architecture for efficiently processing raw audio samples.
Stars: ✭ 764 (+70.54%)
Mutual labels:  speech-recognition, speech-processing
speechrec
a simple speech recognition app using the Web Speech API Interfaces
Stars: ✭ 18 (-95.98%)
Mutual labels:  speech-recognition, speech-processing
spokestack-ios
Spokestack: give your iOS app a voice interface!
Stars: ✭ 27 (-93.97%)
Mutual labels:  speech-recognition, speech-processing

uSpeech library

WARNING!!! DO NOT USE THIS BRANCH. IT IS FOR EXPERIMENTATION. DOWNLOAD IT FROM THE RELEASES PAGE!!!!

Also I do not have the time to maintain this library and given the price of Raspberry Pis you probably want to use them for speech recognition instead of an arduino.

The uSpeech library provides an interface for voice recognition using the Arduino. It currently produces phonemes, often the library will produce junk phonemes. Please bare with it for the time being. A noise removal function is underway.

Minimum Requirements

The library is quite intensive on the processor. Each sample collection takes about 3.2 milliseconds so pay close attention to the time. The library has been tested on the Arduino Uno (ATMega32). Each signal object uses up 160bytes. No real time scheduler should be used with this.

Features

  • Letter based recognition
  • Small memory footprint
  • Arduino Compatible
  • Up to 80% accuracy with words
  • Novel algorithm based on simple calculus
  • Plugs directly into an analogRead() port

Documentation

Head over to the wiki and you will find most of the documentation required.

Algorithm

The library utilizes a special algorithm to enable speech detection. First the complexity of the signal is determined by taking the absolute derivative of the signal multiplying it by a fixed point saclar and then dividing it by the absolute integral of the signal. Consonants (other than R,L,N and M) have a value above 40 and vowels have a value below 40. Consonants, they can be divided into frictaves and plosives. Plosives are like p or b whereas frictaves are like s or z. Generally each band of the complexity coeficient (abs derivative over abs integral) can be matched to a small set of frictaves and plosives. The signal determines if it is a plosive or a frictave by watching the length of the utterance (plosives occur over short periods while frictaves over long). Finally the most appropriate character is chosen.

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].