deepgttDeepGTT: Learning Travel Time Distributions with Deep Generative Model
Stars: ✭ 30 (-92.42%)
char-VAEInspired by the neural style algorithm in the computer vision field, we propose a high-level language model with the aim of adapting the linguistic style.
Stars: ✭ 18 (-95.45%)
probabilistic nlgTensorflow Implementation of Stochastic Wasserstein Autoencoder for Probabilistic Sentence Generation (NAACL 2019).
Stars: ✭ 28 (-92.93%)
MIDI-VAENo description or website provided.
Stars: ✭ 56 (-85.86%)
VAE-Gumbel-SoftmaxAn implementation of a Variational-Autoencoder using the Gumbel-Softmax reparametrization trick in TensorFlow (tested on r1.5 CPU and GPU) in ICLR 2017.
Stars: ✭ 66 (-83.33%)
tensorflow-mnist-AAETensorflow implementation of adversarial auto-encoder for MNIST
Stars: ✭ 86 (-78.28%)
benchmark VAEUnifying Variational Autoencoder (VAE) implementations in Pytorch (NeurIPS 2022)
Stars: ✭ 1,211 (+205.81%)
nvaeAn unofficial toy implementation for NVAE 《A Deep Hierarchical Variational Autoencoder》
Stars: ✭ 83 (-79.04%)
Vae Cvae MnistVariational Autoencoder and Conditional Variational Autoencoder on MNIST in PyTorch
Stars: ✭ 229 (-42.17%)
contiguous-succotashRecurrent Variational Autoencoder with Dilated Convolutions that generates sequential data implemented in pytorch
Stars: ✭ 71 (-82.07%)
pyroVEDInvariant representation learning from imaging and spectral data
Stars: ✭ 23 (-94.19%)
S Vae PytorchPytorch implementation of Hyperspherical Variational Auto-Encoders
Stars: ✭ 255 (-35.61%)
molecular-VAEImplementation of the paper - Automatic chemical design using a data-driven continuous representation of molecules
Stars: ✭ 36 (-90.91%)
concept-based-xaiLibrary implementing state-of-the-art Concept-based and Disentanglement Learning methods for Explainable AI
Stars: ✭ 41 (-89.65%)
Neural OdeJupyter notebook with Pytorch implementation of Neural Ordinary Differential Equations
Stars: ✭ 335 (-15.4%)
InpaintNetCode accompanying ISMIR'19 paper titled "Learning to Traverse Latent Spaces for Musical Score Inpaintning"
Stars: ✭ 48 (-87.88%)
continuous BernoulliThere are C language computer programs about the simulator, transformation, and test statistic of continuous Bernoulli distribution. More than that, the book contains continuous Binomial distribution and continuous Trinomial distribution.
Stars: ✭ 22 (-94.44%)
Fun-with-MNISTPlaying with MNIST. Machine Learning. Generative Models.
Stars: ✭ 23 (-94.19%)
disent🧶 Modular VAE disentanglement framework for python built with PyTorch Lightning ▸ Including metrics and datasets ▸ With strongly supervised, weakly supervised and unsupervised methods ▸ Easily configured and run with Hydra config ▸ Inspired by disentanglement_lib
Stars: ✭ 41 (-89.65%)
Pytorch-RL-CPPA Repository with C++ implementations of Reinforcement Learning Algorithms (Pytorch)
Stars: ✭ 73 (-81.57%)
Advanced Models여러가지 유명한 신경망 모델들을 제공합니다. (DCGAN, VAE, Resnet 등등)
Stars: ✭ 48 (-87.88%)
learning-to-drive-in-5-minutesImplementation of reinforcement learning approach to make a car learn to drive smoothly in minutes
Stars: ✭ 227 (-42.68%)
Generative-ModelRepository for implementation of generative models with Tensorflow 1.x
Stars: ✭ 66 (-83.33%)
DaisyrecA developing recommender system in pytorch. Algorithm: KNN, LFM, SLIM, NeuMF, FM, DeepFM, VAE and so on, which aims to fair comparison for recommender system benchmarks
Stars: ✭ 280 (-29.29%)
sqairImplementation of Sequential Attend, Infer, Repeat (SQAIR)
Stars: ✭ 96 (-75.76%)
vae captioningImplementation of Diverse and Accurate Image Description Using a Variational Auto-Encoder with an Additive Gaussian Encoding Space
Stars: ✭ 58 (-85.35%)
Dsprites DatasetDataset to assess the disentanglement properties of unsupervised learning methods
Stars: ✭ 340 (-14.14%)
generative deep learningGenerative Deep Learning Sessions led by Anugraha Sinha (Machine Learning Tokyo)
Stars: ✭ 24 (-93.94%)
BagelIPCCC 2018: Robust and Unsupervised KPI Anomaly Detection Based on Conditional Variational Autoencoder
Stars: ✭ 45 (-88.64%)
classifying-vae-lstmmusic generation with a classifying variational autoencoder (VAE) and LSTM
Stars: ✭ 27 (-93.18%)
Parallel-Tacotron2PyTorch Implementation of Google's Parallel Tacotron 2: A Non-Autoregressive Neural TTS Model with Differentiable Duration Modeling
Stars: ✭ 149 (-62.37%)
vae-concreteKeras implementation of a Variational Auto Encoder with a Concrete Latent Distribution
Stars: ✭ 51 (-87.12%)
TensorMONKA collection of deep learning models (PyTorch implemtation)
Stars: ✭ 21 (-94.7%)
srVAEVAE with RealNVP prior and Super-Resolution VAE in PyTorch. Code release for https://arxiv.org/abs/2006.05218.
Stars: ✭ 56 (-85.86%)
soft-intro-vae-pytorch[CVPR 2021 Oral] Official PyTorch implementation of Soft-IntroVAE from the paper "Soft-IntroVAE: Analyzing and Improving Introspective Variational Autoencoders"
Stars: ✭ 170 (-57.07%)
style-vaeImplementation of VAE and Style-GAN Architecture Achieving State of the Art Reconstruction
Stars: ✭ 25 (-93.69%)
language-modelsKeras implementations of three language models: character-level RNN, word-level RNN and Sentence VAE (Bowman, Vilnis et al 2016).
Stars: ✭ 39 (-90.15%)
Pytorch rvaeRecurrent Variational Autoencoder that generates sequential data implemented with pytorch
Stars: ✭ 332 (-16.16%)
DeepSSM SysIDOfficial PyTorch implementation of "Deep State Space Models for Nonlinear System Identification", 2020.
Stars: ✭ 62 (-84.34%)
Carla-ppoThis repository hosts a customized PPO based agent for Carla. The goal of this project is to make it easier to interact with and experiment in Carla with reinforcement learning based agents -- this, by wrapping Carla in a gym like environment that can handle custom reward functions, custom debug output, etc.
Stars: ✭ 122 (-69.19%)
VAENAR-TTSPyTorch Implementation of VAENAR-TTS: Variational Auto-Encoder based Non-AutoRegressive Text-to-Speech Synthesis.
Stars: ✭ 66 (-83.33%)
Pytorch RlThis repository contains model-free deep reinforcement learning algorithms implemented in Pytorch
Stars: ✭ 394 (-0.51%)
PycadlPython package with source code from the course "Creative Applications of Deep Learning w/ TensorFlow"
Stars: ✭ 356 (-10.1%)
Beta VaePytorch implementation of β-VAE
Stars: ✭ 326 (-17.68%)
DiffuseVAEA combination of VAE's and Diffusion Models for efficient, controllable and high-fidelity generation from low-dimensional latents
Stars: ✭ 81 (-79.55%)
vqvae-2PyTorch implementation of VQ-VAE-2 from "Generating Diverse High-Fidelity Images with VQ-VAE-2"
Stars: ✭ 65 (-83.59%)