All Projects → sreecodeslayer → ml-am-lm-cmusphinx

sreecodeslayer / ml-am-lm-cmusphinx

Licence: GPL-3.0 License
This is Malayalam Speech Recognition model developed for CMUSphinx. This is now used for Google Summer Code 2016

Programming Languages

perl
6916 projects
python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to ml-am-lm-cmusphinx

TensorFlow-Powered Robot Vision
No description or website provided.
Stars: ✭ 34 (+21.43%)
Mutual labels:  recognition
Word-recognition-EmbedNet-CAB
Code implementation for our ICPR, 2020 paper titled "Improving Word Recognition using Multiple Hypotheses and Deep Embeddings"
Stars: ✭ 19 (-32.14%)
Mutual labels:  recognition
finFindR
An application for dorsal fin image recognition and cataloguing
Stars: ✭ 16 (-42.86%)
Mutual labels:  recognition
tweet-music-recognizer
🎙️ Node.js Bot to identify songs in Twitter videos
Stars: ✭ 57 (+103.57%)
Mutual labels:  recognition
recognition-text
A recognition-text application which recognize any text from an image with 98% to 100% accuracy. Gave support for 92 languages and translate your recognized text to 6 supported languages.
Stars: ✭ 22 (-21.43%)
Mutual labels:  recognition
handwriting.js
A simple API access to the handwriting recognition service of Google IME
Stars: ✭ 89 (+217.86%)
Mutual labels:  recognition
etiketai
Etiketai is an online tool designed to label images, useful for training AI models
Stars: ✭ 63 (+125%)
Mutual labels:  recognition
DLCV2018SPRING
Deep Learning for Computer Vision (CommE 5052) in NTU
Stars: ✭ 38 (+35.71%)
Mutual labels:  recognition
ObjRecPoseEst
Object Detection and 3D Pose Estimation
Stars: ✭ 71 (+153.57%)
Mutual labels:  recognition
RecPlate-lib
基于BP神经网络的车牌识别系统
Stars: ✭ 41 (+46.43%)
Mutual labels:  recognition
ner-d
Python module for Named Entity Recognition (NER) using natural language processing.
Stars: ✭ 14 (-50%)
Mutual labels:  recognition
PocketSphinxUnityDemo
A sample Unity project showing how to use PocketSphinx.
Stars: ✭ 32 (+14.29%)
Mutual labels:  pocketsphinx
fade
A Simulation Framework for Auditory Discrimination Experiments
Stars: ✭ 12 (-57.14%)
Mutual labels:  recognition
farm-animal-tracking
Farm Animal Tracking (FAT)
Stars: ✭ 19 (-32.14%)
Mutual labels:  recognition
CCAligner
🔮 Word by word audio subtitle synchronisation tool and API. Developed under GSoC 2017 with CCExtractor.
Stars: ✭ 131 (+367.86%)
Mutual labels:  pocketsphinx
openface mass compare
An openface script that runs a REST server. Posted images are compared against a large dataset, and the most likely match is returned. Works with https://hub.docker.com/r/uoacer/openface-mass-compare/
Stars: ✭ 22 (-21.43%)
Mutual labels:  recognition
Car-Model-Recognition
Car Model Recognition project
Stars: ✭ 53 (+89.29%)
Mutual labels:  recognition
getting-started
List of ideas for getting started with TimVideos projects
Stars: ✭ 50 (+78.57%)
Mutual labels:  gsoc-2016
nmap vscan
nmap service and application version detection (without nmap installation)
Stars: ✭ 105 (+275%)
Mutual labels:  recognition
keras cv attention models
Keras/Tensorflow attention models including beit,botnet,CMT,CoaT,CoAtNet,convnext,cotnet,davit,efficientdet,efficientnet,fbnet,gmlp,halonet,lcnet,levit,mlp-mixer,mobilevit,nfnets,regnet,resmlp,resnest,resnext,resnetd,swin,tinynet,uniformer,volo,wavemlp,yolor,yolox
Stars: ✭ 159 (+467.86%)
Mutual labels:  recognition

ml-am-lm-cmusphinx

Instructions to use this Model for demo purpose ( I recommand using Unix-like enivironment ).

Firstly download the latest libraries needed to run the recognition:

  1. SphinxBase

  2. PocketSphinx

  3. SphinxTrain

  4. Sphinx4

For more details head over to CMUSphinx Download

Once you have downloaded, extracted to their corresponding folder, install them using:

In a unix-like environment (such as linux, solaris etc):

  • if you downloaded directly from the CVS repository, you need to do this at least once to generate the "configure" file:
$ ./autogen.sh
  • if you downloaded the release version, or ran "autogen.sh" at least once, then compile and install:
 $ ./configure
 $ make clean all
 $ make check
 $ sudo make install

Now, download the zip of this repository, extract and open terminal inside the root folder.

Connect the microphone and use the command below to run the recognition. I cannot assure accuracy as of yet as this a trail attempt towards building a more spanned model.

pocketsphinx_continuous -hmm ./ -lm samsaaram.arpa -dict samsaaram.dic -inmic yes | tee ml_terminal_output_export.txt

####Note:

The installation of libraries can throw many errors depending on the various dependencies of autogen , configure , make . Make sure to patiently resolve those to have a successful installation. Also make sure to set the path variables in the environment.

Audio driver package(s) (osspd generally) of your system might need updation while launching the command :

pocketsphinx_continuous

Try this and all should probably run fine after.

sudo apt-get update
sudo apt-get install osspd

To contribute

  1. Fork this repository.
  2. Record* the sentences^
  3. Commit and make a Pull Request.

####Note:

Record* - To record, use Audacity , set Project Rate = 16000Hz, Default Sample Format as 16bit, and while saving, use WAV, PCM 16bit option

sentences^ - The sentences [file](/Further development files/hugu+interstellar+queen - sentences.txt) can be found inside the file "hugu+interstellar+queen - sentences.txt" under Further Development.

Please contact me before you start recording.

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].