All Projects → ashual → Scene_generation

ashual / Scene_generation

Licence: apache-2.0
A PyTorch implementation of the paper: Specifying Object Attributes and Relations in Interactive Scene Generation

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to Scene generation

Ganspace
Discovering Interpretable GAN Controls [NeurIPS 2020]
Stars: ✭ 1,224 (+674.68%)
Mutual labels:  image-generation
Icface
ICface: Interpretable and Controllable Face Reenactment Using GANs
Stars: ✭ 122 (-22.78%)
Mutual labels:  image-generation
Unetgan
Official Implementation of the paper "A U-Net Based Discriminator for Generative Adversarial Networks" (CVPR 2020)
Stars: ✭ 139 (-12.03%)
Mutual labels:  image-generation
Vue Pwa Asset Generator
PWA asset generator perfect with VueJS framework (but useful for all PWA!)
Stars: ✭ 97 (-38.61%)
Mutual labels:  image-generation
Pwa Asset Generator
Automates PWA asset generation and image declaration. Automatically generates icon and splash screen images, favicons and mstile images. Updates manifest.json and index.html files with the generated images according to Web App Manifest specs and Apple Human Interface guidelines.
Stars: ✭ 1,787 (+1031.01%)
Mutual labels:  image-generation
Cyclegan
Software that can generate photos from paintings, turn horses into zebras, perform style transfer, and more.
Stars: ✭ 10,933 (+6819.62%)
Mutual labels:  image-generation
Semantic Object Accuracy For Generative Text To Image Synthesis
Code for "Semantic Object Accuracy for Generative Text-to-Image Synthesis" (TPAMI 2020)
Stars: ✭ 78 (-50.63%)
Mutual labels:  image-generation
Vae Lagging Encoder
PyTorch implementation of "Lagging Inference Networks and Posterior Collapse in Variational Autoencoders" (ICLR 2019)
Stars: ✭ 153 (-3.16%)
Mutual labels:  image-generation
Pytorch Generative
Easy generative modeling in PyTorch.
Stars: ✭ 112 (-29.11%)
Mutual labels:  image-generation
Fq Gan
Official implementation of FQ-GAN
Stars: ✭ 137 (-13.29%)
Mutual labels:  image-generation
Lggan
[CVPR 2020] Local Class-Specific and Global Image-Level Generative Adversarial Networks for Semantic-Guided Scene Generation
Stars: ✭ 97 (-38.61%)
Mutual labels:  image-generation
Neural Doodle
Turn your two-bit doodles into fine artworks with deep neural networks, generate seamless textures from photos, transfer style from one image to another, perform example-based upscaling, but wait... there's more! (An implementation of Semantic Style Transfer.)
Stars: ✭ 9,680 (+6026.58%)
Mutual labels:  image-generation
Oneshottranslation
Pytorch implementation of "One-Shot Unsupervised Cross Domain Translation" NIPS 2018
Stars: ✭ 135 (-14.56%)
Mutual labels:  image-generation
Vae For Image Generation
Implemented Variational Autoencoder generative model in Keras for image generation and its latent space visualization on MNIST and CIFAR10 datasets
Stars: ✭ 87 (-44.94%)
Mutual labels:  image-generation
Focal Frequency Loss
Focal Frequency Loss for Generative Models
Stars: ✭ 141 (-10.76%)
Mutual labels:  image-generation
Generating Devanagari Using Draw
PyTorch implementation of DRAW: A Recurrent Neural Network For Image Generation trained on Devanagari dataset.
Stars: ✭ 82 (-48.1%)
Mutual labels:  image-generation
Mlds2018spring
Machine Learning and having it Deep and Structured (MLDS) in 2018 spring
Stars: ✭ 124 (-21.52%)
Mutual labels:  image-generation
Mmediting
OpenMMLab Image and Video Editing Toolbox
Stars: ✭ 2,618 (+1556.96%)
Mutual labels:  image-generation
Tsit
[ECCV 2020 Spotlight] A Simple and Versatile Framework for Image-to-Image Translation
Stars: ✭ 141 (-10.76%)
Mutual labels:  image-generation
Gesturegan
[ACM MM 2018 Oral] GestureGAN for Hand Gesture-to-Gesture Translation in the Wild
Stars: ✭ 136 (-13.92%)
Mutual labels:  image-generation

Specifying Object Attributes and Relations in Interactive Scene Generation

A PyTorch implementation of the paper Specifying Object Attributes and Relations in Interactive Scene Generation

Paper

Specifying Object Attributes and Relations in Interactive Scene Generation
Oron Ashual1, Lior Wolf1,2
1 Tel-Aviv University, 2 Facebook AI Research
The IEEE International Conference on Computer Vision (ICCV), 2019, (Oral)

Network Architechture

Youtube

paper_video

Usage

1. Create a virtual environment (optional)

All code was developed and tested on Ubuntu 18.04 with Python 3.6 (Anaconda) and PyTorch 1.0.

conda create -n scene_generation python=3.7
conda activate scene_generation

2. Clone the repository

cd ~
git clone https://github.com/ashual/scene_generation.git
cd scene_generation

3. Install dependencies

conda install --file requirements.txt -c conda-forge -c pytorch
  • install a PyTorch version which will fit your CUDA TOOLKIT

4. Install COCO API

Note: we didn't train our models with COCO panoptic dataset, the coco_panoptic.py code is for the sake of the community only.

cd ~
git clone https://github.com/cocodataset/cocoapi.git
cd cocoapi/PythonAPI/
python setup.py install
cd ~/scene_generation

5. Train

$ python train.py

6. Encode the Appearance attributes

python scripts/encode_features --checkpoint TRAINED_MODEL_CHECKPOINT

7. Sample Images

python scripts/sample_images.py --checkpoint TRAINED_MODEL_CHECKPOINT --batch_size 32 --output_dir OUTPUT_DIR 

8. or Download trained models

Download these files into models/

9. Play with the GUI

The GUI was built as POC. Use it at your own risk:

python scripts/gui/simple-server.py --checkpoint YOUR_MODEL_CHECKPOINT --output_dir [DIR_NAME] --draw_scene_graphs 0

10. Results

Results were measured by sample images from the validation set and then running these 3 official scripts:

  1. FID - https://github.com/bioinf-jku/TTUR (Tensorflow implementation)
  2. Inception - https://github.com/openai/improved-gan/blob/master/inception_score/model.py (Tensorflow implementation)
  3. Diversity - https://github.com/richzhang/PerceptualSimilarity (Pytorch implementation)
  4. Accuracy - Training code is attached train_accuracy_net.py. A trained model is provided. Adding the argument --accuracy_model_path MODEL_PATH will output the accuracy of the objects.

Reproduce the comparison figure (Figure 3.)

Run this command

$ python scripts/sample_images.py --checkpoint TRAINED_MODEL_CHECKPOINT --output_dir OUTPUT_DIR

with these arguments:

  • (c) - Ground truth layout: --use_gt_boxes 1 --use_gt_masks 1
  • (d) - Ground truth location attributes: --use_gt_attr 1
  • (e) - Ground truth appearance attributes: --use_gt_textures 1
  • (f) - Scene Graph only - No extra attributes needed

Citation

If you find this code useful in your research then please cite

@InProceedings{Ashual_2019_ICCV,
    author = {Ashual, Oron and Wolf, Lior},
    title = {Specifying Object Attributes and Relations in Interactive Scene Generation},
    booktitle = {The IEEE International Conference on Computer Vision (ICCV)},
    month = {October},
    year = {2019}
}

Acknowledgement

Our project borrows some source files from sg2im. We thank the authors.

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].