pyroVEDInvariant representation learning from imaging and spectral data
Stars: ✭ 23 (-64.62%)
Awesome VaesA curated list of awesome work on VAEs, disentanglement, representation learning, and generative models.
Stars: ✭ 418 (+543.08%)
BagelIPCCC 2018: Robust and Unsupervised KPI Anomaly Detection Based on Conditional Variational Autoencoder
Stars: ✭ 45 (-30.77%)
Cada Vae PytorchOfficial implementation of the paper "Generalized Zero- and Few-Shot Learning via Aligned Variational Autoencoders" (CVPR 2019)
Stars: ✭ 198 (+204.62%)
VAE-Gumbel-SoftmaxAn implementation of a Variational-Autoencoder using the Gumbel-Softmax reparametrization trick in TensorFlow (tested on r1.5 CPU and GPU) in ICLR 2017.
Stars: ✭ 66 (+1.54%)
Deep Learning With PythonExample projects I completed to understand Deep Learning techniques with Tensorflow. Please note that I do no longer maintain this repository.
Stars: ✭ 134 (+106.15%)
Disentangling VaeExperiments for understanding disentanglement in VAE latent representations
Stars: ✭ 398 (+512.31%)
SmrtHandle class imbalance intelligently by using variational auto-encoders to generate synthetic observations of your minority class.
Stars: ✭ 102 (+56.92%)
classifying-vae-lstmmusic generation with a classifying variational autoencoder (VAE) and LSTM
Stars: ✭ 27 (-58.46%)
benchmark VAEUnifying Variational Autoencoder (VAE) implementations in Pytorch (NeurIPS 2022)
Stars: ✭ 1,211 (+1763.08%)
MIDI-VAENo description or website provided.
Stars: ✭ 56 (-13.85%)
S Vae TfTensorflow implementation of Hyperspherical Variational Auto-Encoders
Stars: ✭ 198 (+204.62%)
Vae For Image GenerationImplemented Variational Autoencoder generative model in Keras for image generation and its latent space visualization on MNIST and CIFAR10 datasets
Stars: ✭ 87 (+33.85%)
Vae TensorflowA Tensorflow implementation of a Variational Autoencoder for the deep learning course at the University of Southern California (USC).
Stars: ✭ 117 (+80%)
MojitalkCode for "MojiTalk: Generating Emotional Responses at Scale" https://arxiv.org/abs/1711.04090
Stars: ✭ 107 (+64.62%)
continuous BernoulliThere are C language computer programs about the simulator, transformation, and test statistic of continuous Bernoulli distribution. More than that, the book contains continuous Binomial distribution and continuous Trinomial distribution.
Stars: ✭ 22 (-66.15%)
S Vae PytorchPytorch implementation of Hyperspherical Variational Auto-Encoders
Stars: ✭ 255 (+292.31%)
srVAEVAE with RealNVP prior and Super-Resolution VAE in PyTorch. Code release for https://arxiv.org/abs/2006.05218.
Stars: ✭ 56 (-13.85%)
Vae protein functionProtein function prediction using a variational autoencoder
Stars: ✭ 57 (-12.31%)
Tensorflow Mnist VaeTensorflow implementation of variational auto-encoder for MNIST
Stars: ✭ 422 (+549.23%)
Pytorch VaeA CNN Variational Autoencoder (CNN-VAE) implemented in PyTorch
Stars: ✭ 181 (+178.46%)
vae-concreteKeras implementation of a Variational Auto Encoder with a Concrete Latent Distribution
Stars: ✭ 51 (-21.54%)
Vae Cvae MnistVariational Autoencoder and Conditional Variational Autoencoder on MNIST in PyTorch
Stars: ✭ 229 (+252.31%)
Tensorflow Mnist CvaeTensorflow implementation of conditional variational auto-encoder for MNIST
Stars: ✭ 139 (+113.85%)
Variational AutoencoderVariational autoencoder implemented in tensorflow and pytorch (including inverse autoregressive flow)
Stars: ✭ 807 (+1141.54%)
Pytorch RlThis repository contains model-free deep reinforcement learning algorithms implemented in Pytorch
Stars: ✭ 394 (+506.15%)
soft-intro-vae-pytorch[CVPR 2021 Oral] Official PyTorch implementation of Soft-IntroVAE from the paper "Soft-IntroVAE: Analyzing and Improving Introspective Variational Autoencoders"
Stars: ✭ 170 (+161.54%)
Fun-with-MNISTPlaying with MNIST. Machine Learning. Generative Models.
Stars: ✭ 23 (-64.62%)
OptimusOptimus: the first large-scale pre-trained VAE language model
Stars: ✭ 180 (+176.92%)
FactorvaePytorch implementation of FactorVAE proposed in Disentangling by Factorising(http://arxiv.org/abs/1802.05983)
Stars: ✭ 176 (+170.77%)
InpaintNetCode accompanying ISMIR'19 paper titled "Learning to Traverse Latent Spaces for Musical Score Inpaintning"
Stars: ✭ 48 (-26.15%)
language-modelsKeras implementations of three language models: character-level RNN, word-level RNN and Sentence VAE (Bowman, Vilnis et al 2016).
Stars: ✭ 39 (-40%)
Vae Lagging EncoderPyTorch implementation of "Lagging Inference Networks and Posterior Collapse in Variational Autoencoders" (ICLR 2019)
Stars: ✭ 153 (+135.38%)
Pytorch VaeA Collection of Variational Autoencoders (VAE) in PyTorch.
Stars: ✭ 2,704 (+4060%)
deep-blueberryIf you've always wanted to learn about deep-learning but don't know where to start, then you might have stumbled upon the right place!
Stars: ✭ 17 (-73.85%)
normalizing-flowsPyTorch implementation of normalizing flow models
Stars: ✭ 271 (+316.92%)
DeepSSM SysIDOfficial PyTorch implementation of "Deep State Space Models for Nonlinear System Identification", 2020.
Stars: ✭ 62 (-4.62%)
Beat BlenderBlend beats using machine learning to create music in a fun new way.
Stars: ✭ 147 (+126.15%)
Vae SeqVariational Auto-Encoders in a Sequential Setting.
Stars: ✭ 145 (+123.08%)
Vmf vae nlpCode for EMNLP18 paper "Spherical Latent Spaces for Stable Variational Autoencoders"
Stars: ✭ 140 (+115.38%)
SIVIUsing neural network to build expressive hierarchical distribution; A variational method to accurately estimate posterior uncertainty; A fast and general method for Bayesian inference. (ICML 2018)
Stars: ✭ 49 (-24.62%)
Srl ZooState Representation Learning (SRL) zoo with PyTorch - Part of S-RL Toolbox
Stars: ✭ 125 (+92.31%)
svae cf[ WSDM '19 ] Sequential Variational Autoencoders for Collaborative Filtering
Stars: ✭ 38 (-41.54%)
O GanO-GAN: Extremely Concise Approach for Auto-Encoding Generative Adversarial Networks
Stars: ✭ 117 (+80%)
Pytorch cppDeep Learning sample programs using PyTorch in C++
Stars: ✭ 114 (+75.38%)
Tf VqvaeTensorflow Implementation of the paper [Neural Discrete Representation Learning](https://arxiv.org/abs/1711.00937) (VQ-VAE).
Stars: ✭ 226 (+247.69%)
Numpy MlMachine learning, in numpy
Stars: ✭ 11,100 (+16976.92%)
VAE-Latent-Space-ExplorerInteractive exploration of MNIST variational autoencoder latent space with React and tensorflow.js.
Stars: ✭ 30 (-53.85%)
Vq VaeMinimalist implementation of VQ-VAE in Pytorch
Stars: ✭ 224 (+244.62%)