All Projects → manthanthakker → speakerIdentificationNeuralNetworks

manthanthakker / speakerIdentificationNeuralNetworks

Licence: MIT License
⇨ The Speaker Recognition System consists of two phases, Feature Extraction and Recognition. ⇨ In the Extraction phase, the Speaker's voice is recorded and typical number of features are extracted to form a model. ⇨ During the Recognition phase, a speech sample is compared against a previously created voice print stored in the database. ⇨ The hi…

Programming Languages

matlab
3953 projects
python
139335 projects - #7 most used programming language
Jupyter Notebook
11667 projects
Mathematica
289 projects

Projects that are alternatives of or similar to speakerIdentificationNeuralNetworks

Huawei-Challenge-Speaker-Identification
Trained speaker embedding deep learning models and evaluation pipelines in pytorch and tesorflow for speaker recognition.
Stars: ✭ 34 (+30.77%)
Mutual labels:  speaker-recognition, speaker-identification
dropclass speaker
DropClass and DropAdapt - repository for the paper accepted to Speaker Odyssey 2020
Stars: ✭ 20 (-23.08%)
Mutual labels:  speaker-recognition, speaker-identification
wavenet-classifier
Keras Implementation of Deepmind's WaveNet for Supervised Learning Tasks
Stars: ✭ 54 (+107.69%)
Mutual labels:  speaker-recognition, speaker-identification
Speaker-Recognition
This repo contains my attempt to create a Speaker Recognition and Verification system using SideKit-1.3.1
Stars: ✭ 94 (+261.54%)
Mutual labels:  speaker-recognition, speaker-identification
KaldiBasedSpeakerVerification
Kaldi based speaker verification
Stars: ✭ 43 (+65.38%)
Mutual labels:  speaker-recognition, speaker-identification
GE2E-Loss
Pytorch implementation of Generalized End-to-End Loss for speaker verification
Stars: ✭ 72 (+176.92%)
Mutual labels:  speaker-recognition, speaker-identification
meta-SR
Pytorch implementation of Meta-Learning for Short Utterance Speaker Recognition with Imbalance Length Pairs (Interspeech, 2020)
Stars: ✭ 58 (+123.08%)
Mutual labels:  speaker-recognition
2018-dlsl
UPC Deep Learning for Speech and Language 2018
Stars: ✭ 18 (-30.77%)
Mutual labels:  speaker-identification
Voiceprint-recognition-Speaker-recognition
It is a complete project of voiceprint recognition or speaker recognition.
Stars: ✭ 82 (+215.38%)
Mutual labels:  speaker-recognition
Piwho
Speaker recognition library based on MARF for raspberry pi and other SBCs.
Stars: ✭ 50 (+92.31%)
Mutual labels:  speaker-recognition
VoiceprintRecognition-Keras
基于Kersa实现的声纹识别模型
Stars: ✭ 70 (+169.23%)
Mutual labels:  speaker-recognition
kaldi-timit-sre-ivector
Develop speaker recognition model based on i-vector using TIMIT database
Stars: ✭ 17 (-34.62%)
Mutual labels:  speaker-recognition
speaker-recognition-pytorch
Speaker recognition ,Voiceprint recognition
Stars: ✭ 49 (+88.46%)
Mutual labels:  speaker-recognition
Speaker-Identification
A program for automatic speaker identification using deep learning techniques.
Stars: ✭ 84 (+223.08%)
Mutual labels:  speaker-recognition
speaker-recognition-papers
Share some recent speaker recognition papers and their implementations.
Stars: ✭ 92 (+253.85%)
Mutual labels:  speaker-recognition
deepaudio-speaker
neural network based speaker embedder
Stars: ✭ 19 (-26.92%)
Mutual labels:  speaker-recognition
FreeSR
A Free Library for Speaker Recognition (Verification),implemented by ncnn.
Stars: ✭ 21 (-19.23%)
Mutual labels:  speaker-recognition
D-TDNN
PyTorch implementation of Densely Connected Time Delay Neural Network
Stars: ✭ 60 (+130.77%)
Mutual labels:  speaker-recognition
AESRC2020
a deep accent recognition network
Stars: ✭ 35 (+34.62%)
Mutual labels:  speaker-recognition
VoiceprintRecognition-Pytorch
本项目使用了EcapaTdnn模型实现的声纹识别
Stars: ✭ 140 (+438.46%)
Mutual labels:  speaker-recognition

Speaker Identification in Multispeaker Environment using Deep Neural Networks

Abstract

Human beings are capable of performing unfathomable tasks. A human being is able to focus on a single person’s voice in an environment of simultaneous conversations. We have tried to emulate this particular skill through an artificial intelligence system. Our system identifies an audio file as a single or multi-speaker file as the first step and then recognizes the speaker(s). Our approach towards the desired solution was to first conduct pre-processing of the audio (input) file where it is subjected to reduction and silence removal, framing, windowing and DCT calculation, all of which is used to extract its features. Mel Frequency Cepstral Coefficients (MFCC) technique was used for feature extraction. The extracted features are then used to train the system via neural networks using the Error Back Propagation Training Algorithm (EBPTA). One of the many applications of our model is in biometric systems such as telephone banking, authentication and surveillance.

Keywords: Speaker identification, neural network, Multi- Speaker, Mel Frequency Cepstral Coefficients (MFCC).

Research Paper published in Springer Journal.

For more details: download file ResearchPaper.pdf, projectreport

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].