All Projects → Totoketchup → Adaptive Multispeaker Separation

Totoketchup / Adaptive Multispeaker Separation

Licence: mit
Adaptive and Focusing Neural Layers for Multi-Speaker Separation Problem

Projects that are alternatives of or similar to Adaptive Multispeaker Separation

Deeplearning
深度学习入门教程, 优秀文章, Deep Learning Tutorial
Stars: ✭ 6,783 (+16050%)
Mutual labels:  jupyter-notebook, deeplearning
Basic reinforcement learning
An introductory series to Reinforcement Learning (RL) with comprehensive step-by-step tutorials.
Stars: ✭ 826 (+1866.67%)
Mutual labels:  jupyter-notebook, deeplearning
Deeplearning Assignment
深度学习笔记
Stars: ✭ 619 (+1373.81%)
Mutual labels:  jupyter-notebook, deeplearning
Additive Margin Softmax
This is the implementation of paper <Additive Margin Softmax for Face Verification>
Stars: ✭ 464 (+1004.76%)
Mutual labels:  jupyter-notebook, deeplearning
Minecraft Reinforcement Learning
Deep Recurrent Q-Learning vs Deep Q Learning on a simple Partially Observable Markov Decision Process with Minecraft
Stars: ✭ 33 (-21.43%)
Mutual labels:  jupyter-notebook, deeplearning
Introtodeeplearning
Lab Materials for MIT 6.S191: Introduction to Deep Learning
Stars: ✭ 4,955 (+11697.62%)
Mutual labels:  jupyter-notebook, deeplearning
Ai Series
📚 [.md & .ipynb] Series of Artificial Intelligence & Deep Learning, including Mathematics Fundamentals, Python Practices, NLP Application, etc. 💫 人工智能与深度学习实战,数理统计篇 | 机器学习篇 | 深度学习篇 | 自然语言处理篇 | 工具实践 Scikit & Tensoflow & PyTorch 篇 | 行业应用 & 课程笔记
Stars: ✭ 702 (+1571.43%)
Mutual labels:  jupyter-notebook, deeplearning
Pytorch Original Transformer
My implementation of the original transformer model (Vaswani et al.). I've additionally included the playground.py file for visualizing otherwise seemingly hard concepts. Currently included IWSLT pretrained models.
Stars: ✭ 411 (+878.57%)
Mutual labels:  jupyter-notebook, deeplearning
Advanced Gradient Obfuscating
Take further steps in the arms race of adversarial examples with only preprocessing.
Stars: ✭ 28 (-33.33%)
Mutual labels:  jupyter-notebook, deeplearning
Servenet
Service Classification based on Service Description
Stars: ✭ 21 (-50%)
Mutual labels:  jupyter-notebook, deeplearning
Pytorch tutorial
PyTorch Tutorial (1.7)
Stars: ✭ 450 (+971.43%)
Mutual labels:  jupyter-notebook, deeplearning
Coursera Natural Language Processing Specialization
Programming assignments from all courses in the Coursera Natural Language Processing Specialization offered by deeplearning.ai.
Stars: ✭ 39 (-7.14%)
Mutual labels:  jupyter-notebook, deeplearning
Monk object detection
A one-stop repository for low-code easily-installable object detection pipelines.
Stars: ✭ 437 (+940.48%)
Mutual labels:  jupyter-notebook, deeplearning
Monk v1
Monk is a low code Deep Learning tool and a unified wrapper for Computer Vision.
Stars: ✭ 480 (+1042.86%)
Mutual labels:  jupyter-notebook, deeplearning
Deep Learning Resources
由淺入深的深度學習資源 Collection of deep learning materials for everyone
Stars: ✭ 422 (+904.76%)
Mutual labels:  jupyter-notebook, deeplearning
Deeplearningmugenknock
でぃーぷらーにんぐを無限にやってディープラーニングでDeepLearningするための実装CheatSheet
Stars: ✭ 684 (+1528.57%)
Mutual labels:  jupyter-notebook, deeplearning
Portrait Segmentation
Real-time portrait segmentation for mobile devices
Stars: ✭ 358 (+752.38%)
Mutual labels:  jupyter-notebook, deeplearning
Text summurization abstractive methods
Multiple implementations for abstractive text summurization , using google colab
Stars: ✭ 359 (+754.76%)
Mutual labels:  jupyter-notebook, deeplearning
Concise Ipython Notebooks For Deep Learning
Ipython Notebooks for solving problems like classification, segmentation, generation using latest Deep learning algorithms on different publicly available text and image data-sets.
Stars: ✭ 23 (-45.24%)
Mutual labels:  jupyter-notebook, deeplearning
Relativistic Average Gan Keras
The implementation of Relativistic average GAN with Keras
Stars: ✭ 36 (-14.29%)
Mutual labels:  jupyter-notebook, deeplearning

Adaptive and Focus Layer for Multi-Speaker Separation Problem

This repository contains the implementation of an Adaptive Layer and Focus Layer for the Multi-Speaker Separation Problem. The Adaptive Layer consists in a sparse linear AutoEncoder replacing the use of Spectograms and dealing directly with raw audio files. This AutoEncoder is added around the current following state of the art architectures:

  • Deep Clustering Model (with and without enhancing layer)
  • L41 (Magnolia) Model (with and without enhancing layer)

These are compared to traditional STFT approaches with the following architectures:

  • Deep Clustering Model (with and without enhancing layer)
  • L41 (Magnolia) Model (with and without enhancing layer)
  • Dense Model (TODO) with 2 speakers
  • PIT Model (TODO) with 2 speakers

The Focus Layer is in constuction.

Training

Adapt Layer

Pretraining

python -m experiments.training.pretraining --men --women --loss sdr+l2 --separation mask --learning_rate 0.001 --nb_speakers 2 --batch_size 4 --filters 256 --max_pool 256 --beta 0.0 --regularization 0.0 --overlap_coef 1.0 --no_random_picking

Arguments:

Related to Data:
  • --men: Use men voices
  • --women: Use women voices
  • --nb_speakers: Number of mixed speakers
  • --no_random_picking: Regular mixing (Man + Female + Man + Female etc...) instead of random mixing
Related to Hyperparams for the Adapt Layer:

Architecture:

  • --filters: Number of AutoEncoder Bases
  • --max_pool: Max pooling size along time axis

Coefficients:

  • --beta: Coefficient for Sparsity loss : KL divergence
  • --regularization: Coefficient for L2 reg on weights
  • --overlap_coef: Coefficient for the overlapping loss
  • --loss: l2 / sdr / l2+sdr
Related to Hyperparams for Training:
  • --batch_size (int)
  • --learning_rate (float)

Current requirements:

scipy==1.0.0
tqdm==4.19.4
SoundFile==0.9.0.post1
matplotlib==2.1.0
numpy==1.12.0
tensorflow_gpu==1.4.0
librosa==0.5.1
haikunator==2.1.0
h5py==2.7.0
scikit_learn==0.19.1
tensorflow==1.5.0rc1
Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].