All Projects → iamyuanchung → VQ-APC

iamyuanchung / VQ-APC

Licence: other
Vector Quantized Autoregressive Predictive Coding (VQ-APC)

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to VQ-APC

awesome-graph-self-supervised-learning
Awesome Graph Self-Supervised Learning
Stars: ✭ 805 (+2267.65%)
Mutual labels:  representation-learning, unsupervised-learning, self-supervised-learning, pre-training
Transferlearning
Transfer learning / domain adaptation / domain generalization / multi-task learning etc. Papers, codes, datasets, applications, tutorials.-迁移学习
Stars: ✭ 8,481 (+24844.12%)
Mutual labels:  representation-learning, unsupervised-learning, self-supervised-learning
Simclr
SimCLRv2 - Big Self-Supervised Models are Strong Semi-Supervised Learners
Stars: ✭ 2,720 (+7900%)
Mutual labels:  representation-learning, unsupervised-learning, self-supervised-learning
awesome-contrastive-self-supervised-learning
A comprehensive list of awesome contrastive self-supervised learning papers.
Stars: ✭ 748 (+2100%)
Mutual labels:  representation-learning, unsupervised-learning, self-supervised-learning
Revisiting-Contrastive-SSL
Revisiting Contrastive Methods for Unsupervised Learning of Visual Representations. [NeurIPS 2021]
Stars: ✭ 81 (+138.24%)
Mutual labels:  representation-learning, unsupervised-learning, self-supervised-learning
temporal-ssl
Video Representation Learning by Recognizing Temporal Transformations. In ECCV, 2020.
Stars: ✭ 46 (+35.29%)
Mutual labels:  unsupervised-learning, self-supervised-learning
Unsupervised Classification
SCAN: Learning to Classify Images without Labels (ECCV 2020), incl. SimCLR.
Stars: ✭ 605 (+1679.41%)
Mutual labels:  representation-learning, unsupervised-learning
Bagofconcepts
Python implementation of bag-of-concepts
Stars: ✭ 18 (-47.06%)
Mutual labels:  representation-learning, unsupervised-learning
Autoregressive Predictive Coding
Autoregressive Predictive Coding: An unsupervised autoregressive model for speech representation learning
Stars: ✭ 138 (+305.88%)
Mutual labels:  representation-learning, unsupervised-learning
Contrastive Predictive Coding
Keras implementation of Representation Learning with Contrastive Predictive Coding
Stars: ✭ 369 (+985.29%)
Mutual labels:  representation-learning, unsupervised-learning
Self Supervised Learning Overview
📜 Self-Supervised Learning from Images: Up-to-date reading list.
Stars: ✭ 73 (+114.71%)
Mutual labels:  representation-learning, unsupervised-learning
naru
Neural Relation Understanding: neural cardinality estimators for tabular data
Stars: ✭ 76 (+123.53%)
Mutual labels:  unsupervised-learning, self-supervised-learning
Lemniscate.pytorch
Unsupervised Feature Learning via Non-parametric Instance Discrimination
Stars: ✭ 532 (+1464.71%)
Mutual labels:  representation-learning, unsupervised-learning
Awesome Vaes
A curated list of awesome work on VAEs, disentanglement, representation learning, and generative models.
Stars: ✭ 418 (+1129.41%)
Mutual labels:  representation-learning, unsupervised-learning
Simclr
PyTorch implementation of SimCLR: A Simple Framework for Contrastive Learning of Visual Representations
Stars: ✭ 750 (+2105.88%)
Mutual labels:  representation-learning, unsupervised-learning
Disentangling Vae
Experiments for understanding disentanglement in VAE latent representations
Stars: ✭ 398 (+1070.59%)
Mutual labels:  representation-learning, unsupervised-learning
Pointglr
Global-Local Bidirectional Reasoning for Unsupervised Representation Learning of 3D Point Clouds (CVPR 2020)
Stars: ✭ 86 (+152.94%)
Mutual labels:  representation-learning, unsupervised-learning
Pytorch Byol
PyTorch implementation of Bootstrap Your Own Latent: A New Approach to Self-Supervised Learning
Stars: ✭ 213 (+526.47%)
Mutual labels:  representation-learning, unsupervised-learning
Variational Ladder Autoencoder
Implementation of VLAE
Stars: ✭ 196 (+476.47%)
Mutual labels:  representation-learning, unsupervised-learning
Contrastive Predictive Coding Pytorch
Contrastive Predictive Coding for Automatic Speaker Verification
Stars: ✭ 223 (+555.88%)
Mutual labels:  representation-learning, unsupervised-learning

Vector-Quantized Autoregressive Predictive Coding

For an improved version of VQ-APC, please refer to this repository.

This repository contains the official implementation of Vector-Quantized Autoregressive Predictive Coding (VQ-APC).

VQ-APC is an extension of APC, which defines a self-supervised task for learning high-level speech representation from unannotated speech. For dependencies and data preprocessing, please refer to the implementation of APC. After the data are ready, here's an example command to train your own VQ-APC model:

python train_vqapc.py --rnn_num_layers 3 \
                      --rnn_hidden_size 512 \
                      --rnn_dropout 0.1 \
                      --rnn_residual \
                      --codebook_size 128 \
                      --code_dim 512 \
                      --gumbel_temperature 0.5 \
                      --apply_VQ 0 0 1 \
                      --optimizer adam \
                      --batch_size 32  \
                      --learning_rate 0.0001 \
                      --epochs 10 \
                      --n_future 5 \
                      --librispeech_home ./librispeech_data/preprocessed \
                      --train_partition train-clean-360 \
                      --train_sampling 1. \
                      --val_partition dev-clean \
                      --val_sampling 1. \
                      --exp_name my_exp \
                      --store_path ./logs

Argument descriptions are available in train_vqapc.py.

TODOs

  • Add scripts that get the learned codebook(s) (essentially the parameters of the nn.Linear layer used to implement the VQ layers)
  • Add scripts that visualize the code-phone co-occurrence (Figure 3 in the paper)

Reference

Please kindly cite our work if you find this repository useful:

@inproceedings{chung2020vqapc,
  title = {Vector-quantized autoregressive predictive coding},
  autohor = {Chung, Yu-An and Tang, Hao and Glass, James},
  booktitle = {Interspeech},
  year = {2020}
}

Contact

You can reach me out via email. Questions and feedback are welcome.

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].