All Projects → mcbuehler → Seg2Eye

mcbuehler / Seg2Eye

Licence: other
Official implementation of "Content-Consistent Generation of Realistic Eyes with Style", ICCW 2019

Programming Languages

python
139335 projects - #7 most used programming language
shell
77523 projects

Projects that are alternatives of or similar to Seg2Eye

Fcos
FCOS: Fully Convolutional One-Stage Object Detection (ICCV'19)
Stars: ✭ 2,839 (+10819.23%)
Mutual labels:  iccv2019
WGAN-GP-TensorFlow
TensorFlow implementations of Wasserstein GAN with Gradient Penalty (WGAN-GP), Least Squares GAN (LSGAN), GANs with the hinge loss.
Stars: ✭ 42 (+61.54%)
Mutual labels:  image-synthesis
gan-ensembling
Invert and perturb GAN images for test-time ensembling
Stars: ✭ 93 (+257.69%)
Mutual labels:  image-synthesis
Sketch2Color-anime-translation
Given a simple anime line-art sketch the model outputs a decent colored anime image using Conditional-Generative Adversarial Networks (C-GANs) concept.
Stars: ✭ 90 (+246.15%)
Mutual labels:  image-synthesis
OpenShapes
A data-driven approach for interactively synthesizing diverse images from semantic label maps.
Stars: ✭ 39 (+50%)
Mutual labels:  image-synthesis
Awesome-Text-to-Image
A Survey on Text-to-Image Generation/Synthesis.
Stars: ✭ 251 (+865.38%)
Mutual labels:  image-synthesis
PytorchSSD-HarDNet
achieves 35.1 mAP while as fast as SSD-vgg16
Stars: ✭ 26 (+0%)
Mutual labels:  iccv2019
SMIT
Pytorch implemenation of Stochastic Multi-Label Image-to-image Translation (SMIT), ICCV Workshops 2019.
Stars: ✭ 37 (+42.31%)
Mutual labels:  iccv2019
ArtGAN
Tensorflow codes for our ICIP-17 and arXiv-1708.09533 works: "ArtGAN: Artwork Synthesis with Conditional Categorial GAN" & "Learning a Generative Adversarial Network for High Resolution Artwork Synthesis "
Stars: ✭ 16 (-38.46%)
Mutual labels:  image-synthesis
TA3N
[ICCV 2019 Oral] TA3N: https://github.com/cmhungsteve/TA3N (Most updated repo)
Stars: ✭ 45 (+73.08%)
Mutual labels:  iccv2019
external-internal-inpainting
[CVPR 2021] EII: Image Inpainting with External-Internal Learning and Monochromic Bottleneck
Stars: ✭ 95 (+265.38%)
Mutual labels:  image-synthesis
text-to-image
Text to Image Synthesis using Generative Adversarial Networks
Stars: ✭ 72 (+176.92%)
Mutual labels:  image-synthesis
pytorch clip guided loss
A simple library that implements CLIP guided loss in PyTorch.
Stars: ✭ 67 (+157.69%)
Mutual labels:  image-synthesis
gpuvmem
GPU Framework for Radio Astronomical Image Synthesis
Stars: ✭ 27 (+3.85%)
Mutual labels:  image-synthesis
NovelViewSynthesis-TensorFlow
A TensorFlow implementation of a simple Novel View Synthesis model on ShapeNet (cars and chairs), KITTI, and Synthia.
Stars: ✭ 47 (+80.77%)
Mutual labels:  image-synthesis
Deep Text Recognition Benchmark
Text recognition (optical character recognition) with deep learning methods.
Stars: ✭ 2,665 (+10150%)
Mutual labels:  iccv2019
Fundus Review
Official website of our paper: Applications of Deep Learning in Fundus Images: A Review. Newly-released datasets and recently-published papers will be updated regularly.
Stars: ✭ 64 (+146.15%)
Mutual labels:  image-synthesis
CoCosNet-v2
CoCosNet v2: Full-Resolution Correspondence Learning for Image Translation
Stars: ✭ 312 (+1100%)
Mutual labels:  image-synthesis
FrameNet
FrameNet: Learning Local Canonical Frames of 3D Surfaces from a Single RGB Image
Stars: ✭ 115 (+342.31%)
Mutual labels:  iccv2019
SLE-GAN
Towards Faster and Stabilized GAN Training for High-fidelity Few-shot Image Synthesis
Stars: ✭ 53 (+103.85%)
Mutual labels:  image-synthesis

License CC BY-NC-SA 4.0 Python 3.6

Content-Consistent Generation of Realistic Eyes with Style

Project page | Paper

This README describes the Seg2Eye model. For the Refiner model, please have a look at this folder.

Abstract

Accurately labeled real-world training data can be scarce, and hence recent works adapt, modify or generate images to boost target datasets. However, retaining relevant details from input data in the generated images is challenging and failure could be critical to the performance on the final task. In this work, we synthesize person-specific eye images that satisfy a given semantic segmentation mask (content), while following the style of a specified person from only a few reference images. We introduce two approaches, (a) one used to win the OpenEDS Synthetic Eye Generation Challenge at ICCVW 2019, and (b) a principled approach to solving the problem involving simultaneous injection of style and content information at multiple scales.

Marcel C. Bühler, Seonwook Park, Shalini De Mello, Xucong Zhang, Otmar Hilliges

VR and AR Workshop at ICCV 2019

Seg2Eye Architecture

For details, please refer to our Paper.

Usage

Dataset Preparation

  1. You need access to the OpenEDS Dataset. Please find more information here.

  2. Unzip all folders and set the base_path to the root folder containing the unpacked subfolders. This folder should also contain the json files with the mappings of file to users (OpenEDS{train,validation,test}userID_mapping_to_images.json).

In 'data/prepare_openeds.py', update the 'base_path = "..."' with the path to the unzipped OpenEDS Dataset. Then run

python data/prepare_openeds.py

This will produce an H5 file that you can use to train or test Seg2Eye models.

Training New Models

Run

python train.py --dataroot PATH_TO_H5_FILE

Please note:

  • This implementation currently does not support multi-GPU training.

Testing

python test.py --dataroot PATH_TO_H5_FILE --name CHECKPOINT_NAME \
    --dataset_key VALIDATION|TEST --load_from_opt_file

Code Structure

  • train.py, test.py: the entry point for training and testing.
  • trainers/pix2pix_trainer.py: harnesses and reports the progress of training.
  • models/pix2pix_model.py: creates the networks, and compute the losses
  • models/networks/: defines the architecture of all models
  • options/: creates option lists using argparse package. More individuals are dynamically added in other files as well. Please see the section below.
  • data/: defines the class for loading and processing data.

License

Copyright (C) 2019 NVIDIA Corporation.

All rights reserved. Licensed under the CC BY-NC-SA 4.0 (Attribution-NonCommercial-ShareAlike 4.0 International)

The code is released for academic research use only.

Citation

If you reference our work, please cite our paper.

@inproceedings{Buehler2019ICCVW,
  author    = {Marcel C. Buehler and Seonwook Park and Shalini De Mello and Xucong Zhang and Otmar Hilliges},
  title     = {Content-Consistent Generation of Realistic Eyes with Style},
  year      = {2019},
  booktitle = {International Conference on Computer Vision Workshops (ICCVW)},
  location  = {Seoul, Korea}
}

Acknowledgments

This repository is a fork of the original SPADE implementation.

This work was supported in part by the ERC Grant OPTINT (StG-2016-717054).

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].