All Projects → bjfu-ai-institute → speaker-recognition-papers

bjfu-ai-institute / speaker-recognition-papers

Licence: other
Share some recent speaker recognition papers and their implementations.

Programming Languages

python
139335 projects - #7 most used programming language
shell
77523 projects

Projects that are alternatives of or similar to speaker-recognition-papers

Speaker-Recognition
This repo contains my attempt to create a Speaker Recognition and Verification system using SideKit-1.3.1
Stars: ✭ 94 (+2.17%)
Mutual labels:  speaker-recognition, speaker-verification
deepaudio-speaker
neural network based speaker embedder
Stars: ✭ 19 (-79.35%)
Mutual labels:  speaker-recognition, speaker-verification
wavenet-classifier
Keras Implementation of Deepmind's WaveNet for Supervised Learning Tasks
Stars: ✭ 54 (-41.3%)
Mutual labels:  speaker-recognition, speaker-verification
Huawei-Challenge-Speaker-Identification
Trained speaker embedding deep learning models and evaluation pipelines in pytorch and tesorflow for speaker recognition.
Stars: ✭ 34 (-63.04%)
Mutual labels:  speaker-recognition, speaker-verification
Speaker-Identification
A program for automatic speaker identification using deep learning techniques.
Stars: ✭ 84 (-8.7%)
Mutual labels:  speaker-recognition, speaker-verification
KaldiBasedSpeakerVerification
Kaldi based speaker verification
Stars: ✭ 43 (-53.26%)
Mutual labels:  speaker-recognition, speaker-verification
bob
Bob is a free signal-processing and machine learning toolbox originally developed by the Biometrics group at Idiap Research Institute, in Switzerland. - Mirrored from https://gitlab.idiap.ch/bob/bob
Stars: ✭ 38 (-58.7%)
Mutual labels:  speaker-recognition, speaker-verification
D-TDNN
PyTorch implementation of Densely Connected Time Delay Neural Network
Stars: ✭ 60 (-34.78%)
Mutual labels:  speaker-recognition, speaker-verification
kaldi-timit-sre-ivector
Develop speaker recognition model based on i-vector using TIMIT database
Stars: ✭ 17 (-81.52%)
Mutual labels:  speaker-recognition, speaker-verification
meta-SR
Pytorch implementation of Meta-Learning for Short Utterance Speaker Recognition with Imbalance Length Pairs (Interspeech, 2020)
Stars: ✭ 58 (-36.96%)
Mutual labels:  speaker-recognition, speaker-verification
dropclass speaker
DropClass and DropAdapt - repository for the paper accepted to Speaker Odyssey 2020
Stars: ✭ 20 (-78.26%)
Mutual labels:  speaker-recognition, speaker-verification
GE2E-Loss
Pytorch implementation of Generalized End-to-End Loss for speaker verification
Stars: ✭ 72 (-21.74%)
Mutual labels:  speaker-recognition, speaker-verification
Plant-Disease-Identification-using-CNN
Plant Disease Identification Using Convulutional Neural Network
Stars: ✭ 89 (-3.26%)
Mutual labels:  paper-implementations
logically
explorations in core.logic
Stars: ✭ 108 (+17.39%)
Mutual labels:  paper-implementations
mrnet
Building an ACL tear detector to spot knee injuries from MRIs with PyTorch (MRNet)
Stars: ✭ 98 (+6.52%)
Mutual labels:  paper-implementations
Voice-ML
MobileNet trained with VoxCeleb dataset and used for voice verification
Stars: ✭ 15 (-83.7%)
Mutual labels:  speaker-verification
CycleGAN
A simple code of CycleGAN which is easy to read is implemented by TensorFlow
Stars: ✭ 21 (-77.17%)
Mutual labels:  paper-implementations
Voiceprint-recognition-Speaker-recognition
It is a complete project of voiceprint recognition or speaker recognition.
Stars: ✭ 82 (-10.87%)
Mutual labels:  speaker-recognition
AutoSpeech
[InterSpeech 2020] "AutoSpeech: Neural Architecture Search for Speaker Recognition" by Shaojin Ding*, Tianlong Chen*, Xinyu Gong, Weiwei Zha, Zhangyang Wang
Stars: ✭ 195 (+111.96%)
Mutual labels:  speaker-recognition
dgcnn
Clean & Documented TF2 implementation of "An end-to-end deep learning architecture for graph classification" (M. Zhang et al., 2018).
Stars: ✭ 21 (-77.17%)
Mutual labels:  paper-implementations

Introduction

These are the slightly modified tensorflow/python implementation of recent speaker recognition papers. Please tell me if it is copyright infringement, I'll delete these paper as soon as I can. Our license only apply to our code these paper is not included. Thx.

The file structure is as follows:

|———pyasv
|
|—————model (folder, contain the model)
|
|—————loss (folder, contain the customized loss function)
|
|—————papers (folder, contain the origin paper of most of method)
|
|—————backend(TODO: folder, contain the method of backend)
|
|———data_manage.py (contain some method to manage data)
|
|———speech_processing.py (contain some method to extractfeature and process audio)
|
|———config.py (settings. e.g. save path, learning rate)

More info: Doc

If you want run these code on your computer, you only need to write code like this:

from pyasv import Config
from pyasv.speech_processing import ext_mfcc_feature
from pyasv.data_manage import DataManage
from pyasv.model.ctdnn import run

config = pyasv.Config(name='my_ctdnn_model',
                    n_speaker=1e3,
                    batch_size=64,
                    n_gpu=2,
                    max_step=100,
                    is_big_dataset=False,
                    url_of_bigdataset_temp_file=None,
                    learning_rate=1e-3,
                    slide_windows=[4, 4]
                    save_path='/home/my_path')
config.save('./my_config_path')

frames, labels = ext_mfcc_feature('data_set_path', config)
train = DataManage(frames, labels, config)

frames, labels = ext_mfcc_feature('data_set_path', config)
validation = DataManage(frames, labels, config)

run(config, train, validation)

TODO

  • Implement papers of ICASSP 2018 & Interspeech 2018.
  • Compare each model on a same dataset.

Implemented papers:

  • L. Li, Z. Tang, D. Wang, T. Zheng, "Deep Speaker Feature Learning for Text-Independent Speaker Verification."
  • L. Li, Z. Tang, D. Wang, T. Zheng, "Full-info Training for Deep Speaker Feature Learning," ICASSP 2018.
  • C. Li, X. Ma, B. Jiang, X. Li, X. Zhang, X. Liu, Y. Cao, A. Kannan, Z. Zhu, "Deep Speaker: an End-to-End Neural Speaker Embedding System."
  • Sergey Novoselov, Oleg Kudashev, Vadim Shchemelinin, Ivan Kremnev, Galina Lavrentyeva, "DEEP CNN BASED FEATURE EXTRACTOR FOR TEXT-PROMPTED SPEAKERRECOGNITION."
Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].