All Projects → changliu00 → cygen

changliu00 / cygen

Licence: MIT license
Codes for CyGen, the novel generative modeling framework proposed in "On the Generative Utility of Cyclic Conditionals" (NeurIPS-21)

Programming Languages

python
139335 projects - #7 most used programming language
shell
77523 projects

Projects that are alternatives of or similar to cygen

Variational Ladder Autoencoder
Implementation of VLAE
Stars: ✭ 196 (+345.45%)
Mutual labels:  generative-model
worlds
Building Virtual Reality Worlds using Three.js
Stars: ✭ 23 (-47.73%)
Mutual labels:  generative-model
eccv16 attr2img
Torch Implemention of ECCV'16 paper: Attribute2Image
Stars: ✭ 93 (+111.36%)
Mutual labels:  generative-model
Triple Gan
See Triple-GAN-V2 in PyTorch: https://github.com/taufikxu/Triple-GAN
Stars: ✭ 203 (+361.36%)
Mutual labels:  generative-model
glico-learning-small-sample
Generative Latent Implicit Conditional Optimization when Learning from Small Sample ICPR 20'
Stars: ✭ 20 (-54.55%)
Mutual labels:  generative-model
trVAE
Conditional out-of-distribution prediction
Stars: ✭ 47 (+6.82%)
Mutual labels:  generative-model
Dragan
A stable algorithm for GAN training
Stars: ✭ 189 (+329.55%)
Mutual labels:  generative-model
auto coding
A basic and simple tool for code auto completion
Stars: ✭ 42 (-4.55%)
Mutual labels:  generative-model
naru
Neural Relation Understanding: neural cardinality estimators for tabular data
Stars: ✭ 76 (+72.73%)
Mutual labels:  generative-model
AC-VRNN
PyTorch code for CVIU paper "AC-VRNN: Attentive Conditional-VRNN for Multi-Future Trajectory Prediction"
Stars: ✭ 21 (-52.27%)
Mutual labels:  generative-model
Tf Vqvae
Tensorflow Implementation of the paper [Neural Discrete Representation Learning](https://arxiv.org/abs/1711.00937) (VQ-VAE).
Stars: ✭ 226 (+413.64%)
Mutual labels:  generative-model
Sgan
Stacked Generative Adversarial Networks
Stars: ✭ 240 (+445.45%)
Mutual labels:  generative-model
caffe-simnets
The SimNets Architecture's Implementation in Caffe
Stars: ✭ 13 (-70.45%)
Mutual labels:  generative-model
Neuralnetworks.thought Experiments
Observations and notes to understand the workings of neural network models and other thought experiments using Tensorflow
Stars: ✭ 199 (+352.27%)
Mutual labels:  generative-model
feed forward vqgan clip
Feed forward VQGAN-CLIP model, where the goal is to eliminate the need for optimizing the latent space of VQGAN for each input prompt
Stars: ✭ 135 (+206.82%)
Mutual labels:  generative-model
Voxel Flow
Video Frame Synthesis using Deep Voxel Flow (ICCV 2017 Oral)
Stars: ✭ 191 (+334.09%)
Mutual labels:  generative-model
InpaintNet
Code accompanying ISMIR'19 paper titled "Learning to Traverse Latent Spaces for Musical Score Inpaintning"
Stars: ✭ 48 (+9.09%)
Mutual labels:  generative-model
MMD-GAN
Improving MMD-GAN training with repulsive loss function
Stars: ✭ 82 (+86.36%)
Mutual labels:  generative-model
texturize
🤖🖌️ Generate photo-realistic textures based on source images. Remix, remake, mashup! Useful if you want to create variations on a theme or elaborate on an existing texture.
Stars: ✭ 495 (+1025%)
Mutual labels:  generative-model
PREREQ-IAAI-19
Inferring Concept Prerequisite Relations from Online Educational Resources (IAAI-19)
Stars: ✭ 22 (-50%)
Mutual labels:  generative-model

On the Generative Utility of Cyclic Conditionals

This repository is the official implementation of "On the Generative Utility of Cyclic Conditionals" (NeurIPS 2021).

Chang Liu <[email protected]>, Haoyue Tang, Tao Qin, Jintao Wang, Tie-Yan Liu.
[Paper & Appendix] [Slides] [Video] [Poster]

Introduction

graphical summary

Whether and how can two conditional models p(x|z) and q(z|x) that form a cycle uniquely determine a joint distribution p(x,z)? We develop a general theory for this question, including criteria for the two conditionals to correspond to a common joint (compatibility) and for such joint to be unique (determinacy). As in generative models we need a generator (decoder/likelihood model) and also an encoder (inference model) for representation, the theory indicates they could already define a generative model p(x,z) without specifying a prior distribution p(z)! We call this novel generative modeling framework as CyGen, and develop methods to achieve the eligibility (compatibility and determinacy) and the usage (fitting and generating data) as a generative model.

This codebase implements these CyGen methods, and various baseline methods. The model architectures are based on the Sylvester flow (Householder version), and the experiment environments/setups follow FFJORD. Authorship is clarified in each file.

Requirements

The code requires python version >= 3.6, and is based on PyTorch. To install requirements:

pip install -r requirements.txt

Usage

Run the run_toy.sh and run_image.sh scripts for the synthetic and real-world (i.e. MNIST and SVHN) experiments. See the commands in the script files or python3 main_[toy|image].py --help for customized usage or hyperparameter tuning.

For the real-world experiments, downstream classification accuracy is evaluated along training. To evaluate the FID score, run the command python3 compute_gen_fid.py --load_dict=<path_to_model.pth>.

Results

CyGen synthetic results

As a trailer, we show the synthetic results here. We see that CyGen achieves both high-quality data generation, and well-separated latent clusters (useful representation). This is due to the removal of a specified prior distribution so that the manifold mismatch and posterior collapse problems are avoided. DAE (denoising auto-encoder) does not need a prior, but its training method hurts determinacy. If pretrained as a VAE (i.e. CyGen(PT)), we see that the knowledge of a centered and centrosymmetric prior is encoded through the conditional models. See the paper for more results.

Citation

@inproceedings{liu2021generative,
  author = {Liu, Chang and Tang, Haoyue and Qin, Tao and Wang, Jintao and Liu, Tie-Yan},
  booktitle = {Advances in Neural Information Processing Systems},
  editor = {M. Ranzato and A. Beygelzimer and Y. Dauphin and P.S. Liang and J. Wortman Vaughan},
  pages = {30242--30256},
  publisher = {Curran Associates, Inc.},
  title = {On the Generative Utility of Cyclic Conditionals},
  url = {https://proceedings.neurips.cc/paper/2021/file/fe04e05fbe48920b8ba90bea2ddfe60b-Paper.pdf},
  volume = {34},
  year = {2021}
}
Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].