All Projects → ubclaunchpad → minutes

ubclaunchpad / minutes

Licence: other
🔭 Speaker diarization via transfer learning

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to minutes

LIUM
Scripts for LIUM SpkDiarization tools
Stars: ✭ 28 (+12%)
Mutual labels:  speech, speaker-diarization
D-TDNN
PyTorch implementation of Densely Connected Time Delay Neural Network
Stars: ✭ 60 (+140%)
Mutual labels:  speech, speaker-diarization
wikipron
Massively multilingual pronunciation mining
Stars: ✭ 167 (+568%)
Mutual labels:  speech
StreamingSpeakerDiarization
Official open source implementation of the paper "Overlap-aware low-latency online speaker diarization based on end-to-end local segmentation"
Stars: ✭ 79 (+216%)
Mutual labels:  speaker-diarization
editts
Official implementation of EdiTTS: Score-based Editing for Controllable Text-to-Speech
Stars: ✭ 74 (+196%)
Mutual labels:  speech
jarvis
Jarvis Home Automation
Stars: ✭ 81 (+224%)
Mutual labels:  speech
Meta-Fine-Tuning
[CVPR 2020 VL3] The repository for meta fine-tuning in cross-domain few-shot learning.
Stars: ✭ 29 (+16%)
Mutual labels:  transfer-learning
tradaboost
Transfer learning algorithm TrAdaboost,coded by python
Stars: ✭ 96 (+284%)
Mutual labels:  transfer-learning
flite-go
Go bindings for Flite (festival-lite)
Stars: ✭ 14 (-44%)
Mutual labels:  speech
ML-Research-Made-Easy
Link of ML papers to their blogs/ supplementary material
Stars: ✭ 25 (+0%)
Mutual labels:  transfer-learning
tt-vae-gan
Timbre transfer with variational autoencoding and cycle-consistent adversarial networks. Able to transfer the timbre of an audio source to that of another.
Stars: ✭ 37 (+48%)
Mutual labels:  speech
revolver
REVOLVER - Repeated Evolution in Cancer
Stars: ✭ 52 (+108%)
Mutual labels:  transfer-learning
ser-with-w2v2
Official implementation of INTERSPEECH 2021 paper 'Emotion Recognition from Speech Using Wav2vec 2.0 Embeddings'
Stars: ✭ 40 (+60%)
Mutual labels:  speech
Deep-Learning
It contains the coursework and the practice I have done while learning Deep Learning.🚀 👨‍💻💥 🚩🌈
Stars: ✭ 21 (-16%)
Mutual labels:  transfer-learning
torch-asg
Auto Segmentation Criterion (ASG) implemented in pytorch
Stars: ✭ 42 (+68%)
Mutual labels:  speech
SHOT-plus
code for our TPAMI 2021 paper "Source Data-absent Unsupervised Domain Adaptation through Hypothesis Transfer and Labeling Transfer"
Stars: ✭ 46 (+84%)
Mutual labels:  transfer-learning
CIFAR10-VGG19-Tensorflow
No description or website provided.
Stars: ✭ 27 (+8%)
Mutual labels:  transfer-learning
ObjectNet
PyTorch implementation of "Pyramid Scene Parsing Network".
Stars: ✭ 15 (-40%)
Mutual labels:  transfer-learning
Filipino-Text-Benchmarks
Open-source benchmark datasets and pretrained transformer models in the Filipino language.
Stars: ✭ 22 (-12%)
Mutual labels:  transfer-learning
GenomeAnalysisModule
Welcome to the website and github repository for the Genome Analysis Module. This website will guide the learning experience for trainees in the UBC MSc Genetic Counselling Training Program, as they embark on a journey to learn about analyzing genomes.
Stars: ✭ 19 (-24%)
Mutual labels:  ubc

minutes

Build Status Coverage Status

Jotting things down, so you don't have to.

Spectrogram

Minutes is a speaker diarisation library. Speaker diarisation is the process of identifying different speakers in an audio segment. It is useful for making transcriptions of conversations meaningful by tagging homogenous sections of the conversation with the appropriate speaker.

For more information about Minutes, and how it works, check out our Medium post!

☝️ Installation

Requires Python 3.6!

# Currently we recommend running using the pipenv shell below.
python setup.py install

🏃 Development

Dependencies are managed using a Pipfile and Pipenv:

pipenv install
pipenv shell

Testing

pytest --cov=minutes -vvv test

Example Usage

from minutes import Speaker, Minutes, Conversation

minutes = Minutes(parent='cnn')

# Create some speakers, add some audio.
s1, s2 = Speaker('s1'), Speaker('s2')
s1.add_audio('path/to/audio1')
s2.add_audio('path/to/audio2')

# Add speakers to the model.
minutes.add_speakers([s1, s2])

# Fit the model.
minutes.fit()

# Collect a new conversation for prediction.
conversation = Conversation('/path/to/conversation.wav')

# Create phrases from the conversation.
phrases = minutes.phrases(conversation)
Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].