All Projects → RahulBhalley → progressive-growing-of-gans.pytorch

RahulBhalley / progressive-growing-of-gans.pytorch

Licence: MIT license
Unofficial PyTorch implementation of "Progressive Growing of GANs for Improved Quality, Stability, and Variation".

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to progressive-growing-of-gans.pytorch

progressive growing of GANs
Pure tensorflow implementation of progressive growing of GANs
Stars: ✭ 31 (-39.22%)
Mutual labels:  generative-adversarial-network, wgan, wasserstein-gan, progressive-gan
Improved-Wasserstein-GAN-application-on-MRI-images
Improved Wasserstein GAN (WGAN-GP) application on medical (MRI) images
Stars: ✭ 23 (-54.9%)
Mutual labels:  generative-adversarial-network, wgan, wasserstein-gan
gan-qp.pytorch
Unofficial PyTorch implementation of "GAN-QP: A Novel GAN Framework without Gradient Vanishing and Lipschitz Constraint"
Stars: ✭ 26 (-49.02%)
Mutual labels:  generative-adversarial-network, optimal-transport
TadGAN
Code for the paper "TadGAN: Time Series Anomaly Detection Using Generative Adversarial Networks"
Stars: ✭ 67 (+31.37%)
Mutual labels:  generative-adversarial-network, wasserstein-gan
Awesome Gans
Awesome Generative Adversarial Networks with tensorflow
Stars: ✭ 585 (+1047.06%)
Mutual labels:  generative-adversarial-network, wgan
GAN-Anime-Characters
Applied several Generative Adversarial Networks (GAN) techniques such as: DCGAN, WGAN and StyleGAN to generate Anime Faces and Handwritten Digits.
Stars: ✭ 43 (-15.69%)
Mutual labels:  generative-adversarial-network, wgan
skip-thought-gan
Generating Text through Adversarial Training(GAN) using Skip-Thought Vectors
Stars: ✭ 44 (-13.73%)
Mutual labels:  generative-adversarial-network, wasserstein-gan
Generative Models
Annotated, understandable, and visually interpretable PyTorch implementations of: VAE, BIRVAE, NSGAN, MMGAN, WGAN, WGANGP, LSGAN, DRAGAN, BEGAN, RaGAN, InfoGAN, fGAN, FisherGAN
Stars: ✭ 438 (+758.82%)
Mutual labels:  generative-adversarial-network, wgan
Pggan Pytorch
🔥🔥 PyTorch implementation of "Progressive growing of GANs (PGGAN)" 🔥🔥
Stars: ✭ 653 (+1180.39%)
Mutual labels:  generative-adversarial-network, tensorboard
Generative adversarial networks 101
Keras implementations of Generative Adversarial Networks. GANs, DCGAN, CGAN, CCGAN, WGAN and LSGAN models with MNIST and CIFAR-10 datasets.
Stars: ✭ 138 (+170.59%)
Mutual labels:  generative-adversarial-network, wgan
Dcgan wgan wgan Gp lsgan sngan rsgan began acgan pggan tensorflow
Implementation of some different variants of GANs by tensorflow, Train the GAN in Google Cloud Colab, DCGAN, WGAN, WGAN-GP, LSGAN, SNGAN, RSGAN, RaSGAN, BEGAN, ACGAN, PGGAN, pix2pix, BigGAN
Stars: ✭ 166 (+225.49%)
Mutual labels:  generative-adversarial-network, wgan
Wasserstein2GenerativeNetworks
PyTorch implementation of "Wasserstein-2 Generative Networks" (ICLR 2021)
Stars: ✭ 38 (-25.49%)
Mutual labels:  generative-adversarial-network, optimal-transport
WGAN-GP-tensorflow
Tensorflow Implementation of Paper "Improved Training of Wasserstein GANs"
Stars: ✭ 23 (-54.9%)
Mutual labels:  generative-adversarial-network, wgan
coursera-gan-specialization
Programming assignments and quizzes from all courses within the GANs specialization offered by deeplearning.ai
Stars: ✭ 277 (+443.14%)
Mutual labels:  generative-adversarial-network, wgan
MMD-GAN
Improving MMD-GAN training with repulsive loss function
Stars: ✭ 82 (+60.78%)
Mutual labels:  discriminator, generative-adversarial-network
Fast Srgan
A Fast Deep Learning Model to Upsample Low Resolution Videos to High Resolution at 30fps
Stars: ✭ 417 (+717.65%)
Mutual labels:  generative-adversarial-network, tensorboard
chainer-wasserstein-gan
Chainer implementation of the Wesserstein GAN
Stars: ✭ 20 (-60.78%)
Mutual labels:  wgan, wasserstein-gan
Cartoongan Tensorflow
Generate your own cartoon-style images with CartoonGAN (CVPR 2018), powered by TensorFlow 2.0 Alpha.
Stars: ✭ 587 (+1050.98%)
Mutual labels:  generative-adversarial-network, tensorboard
Pytorch Generative Model Collections
Collection of generative models in Pytorch version.
Stars: ✭ 2,296 (+4401.96%)
Mutual labels:  generative-adversarial-network, wgan
Deeplearning.ai-GAN-Specialization-Generative-Adversarial-Networks
This repository contains my full work and notes on Deeplearning.ai GAN Specialization (Generative Adversarial Networks)
Stars: ✭ 59 (+15.69%)
Mutual labels:  discriminator, generative-adversarial-network

Progressive Growing of Generative Adversarial Network

This is PyTorch implementation of Progressive Growing GANs. The network is trainable on custom image dataset.

Place your dataset folder inside data folder. The training stats are added to repo folder as the training progresses.

Training Configuration

The network training parameters can be configured with following flags.

General settings

  • --train_data_root - set your data sirectory
  • --random_seed - random seed to reproduce the experiments
  • --n_gpu - multiple GPU training

Training parameters

  • --lr - learning rate
  • --lr_decay - learning rate decay at every resolution transition
  • --eps_drift - coefficient for the drift loss
  • --smoothing - smoothing factor for smoothed generator
  • --nc - number of input channel
  • --nz - input dimension of noise
  • --ngf - feature dimension of final layer of generator
  • --ndf - feature dimension of first layer of discriminator
  • --TICK - 1 tick = 1000 images = (1000/batch_size) iteration
  • --max_resl - 10-->1024, 9-->512, 8-->256
  • --trns_tick - transition tick
  • --stab_tick - stabilization tick

Network structure

  • --flag_wn - use of equalized-learning rate
  • --flag_bn - use of batch-normalization (not recommended)
  • --flag_pixelwise - use of pixelwise normalization for generator
  • --flag_gdrop - use of generalized dropout layer for discriminator
  • --flag_leaky - use of leaky relu instead of relu
  • --flag_tanh - use of tanh at the end of the generator
  • --flag_sigmoid - use of sigmoid at the end of the discriminator
  • --flag_add_noise - add noise to the real image(x)
  • --flag_norm_latent - pixelwise normalization of latent vector (z)
  • --flag_add_drift - add drift loss

Optimizer setting

  • --optimizer - optimizer type
  • --beta1 - beta1 for adam
  • --beta2 - beta2 for adam

Display and save setting

  • --use_tb - enable tensorboard visualization
  • --save_img_every - save images every specified iteration
  • --display_tb_every - display progress every specified iteration

GPU Note

Make sure your machine has CUDA enabled GPU(s) if you want to train on GPUs. Change the --n_gpu flag to positive integral value <= available number of GPUs.

TODO

  • WGAN training methodology

Related Links

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].