All Projects → RenYurui → Global Flow Local Attention

RenYurui / Global Flow Local Attention

Licence: other
The source code for paper "Deep Image Spatial Transformation for Person Image Generation"

Projects that are alternatives of or similar to Global Flow Local Attention

Artistic Style Transfer
Convolutional neural networks for artistic style transfer.
Stars: ✭ 341 (-1.16%)
Mutual labels:  jupyter-notebook
Face Recognition
Deep face recognition with Keras, Dlib and OpenCV
Stars: ✭ 342 (-0.87%)
Mutual labels:  jupyter-notebook
Tusimple Benchmark
Download Datasets and Ground Truths: https://github.com/TuSimple/tusimple-benchmark/issues/3
Stars: ✭ 340 (-1.45%)
Mutual labels:  jupyter-notebook
Numpy neural network
仅使用numpy从头开始实现神经网络,包括反向传播公式推导过程; numpy构建全连接层、卷积层、池化层、Flatten层;以及图像分类案例及精调网络案例等,持续更新中... ...
Stars: ✭ 339 (-1.74%)
Mutual labels:  jupyter-notebook
Azureml Bert
End-to-End recipes for pre-training and fine-tuning BERT using Azure Machine Learning Service
Stars: ✭ 342 (-0.87%)
Mutual labels:  jupyter-notebook
Experiments with python
experiments with python
Stars: ✭ 342 (-0.87%)
Mutual labels:  jupyter-notebook
Kaggle criteo ctr challenge
This is a kaggle challenge project called Display Advertising Challenge by CriteoLabs at 2014.这是2014年由CriteoLabs在kaggle上发起的广告点击率预估挑战项目。
Stars: ✭ 340 (-1.45%)
Mutual labels:  jupyter-notebook
Pineapple
Stars: ✭ 344 (-0.29%)
Mutual labels:  jupyter-notebook
Cs231n Camp
cs231n training camp
Stars: ✭ 342 (-0.87%)
Mutual labels:  jupyter-notebook
Trpo
Trust Region Policy Optimization with TensorFlow and OpenAI Gym
Stars: ✭ 343 (-0.58%)
Mutual labels:  jupyter-notebook
Dsprites Dataset
Dataset to assess the disentanglement properties of unsupervised learning methods
Stars: ✭ 340 (-1.45%)
Mutual labels:  jupyter-notebook
Medmnist
[ISBI'21] MedMNIST Classification Decathlon: A Lightweight AutoML Benchmark for Medical Image Analysis
Stars: ✭ 338 (-2.03%)
Mutual labels:  jupyter-notebook
Covid19 Forecast Hub
Projections of COVID-19, in standardized format
Stars: ✭ 342 (-0.87%)
Mutual labels:  jupyter-notebook
Python Topic Model
Implementation of various topic models
Stars: ✭ 339 (-1.74%)
Mutual labels:  jupyter-notebook
Gumbel Softmax
categorical variational autoencoder using the Gumbel-Softmax estimator
Stars: ✭ 343 (-0.58%)
Mutual labels:  jupyter-notebook
Ml Suite
Getting Started with Xilinx ML Suite
Stars: ✭ 340 (-1.45%)
Mutual labels:  jupyter-notebook
Distiller
Neural Network Distiller by Intel AI Lab: a Python package for neural network compression research. https://intellabs.github.io/distiller
Stars: ✭ 3,760 (+989.86%)
Mutual labels:  jupyter-notebook
Deltapy
DeltaPy - Tabular Data Augmentation (by @firmai)
Stars: ✭ 344 (-0.29%)
Mutual labels:  jupyter-notebook
Pytorchnethub
项目注释+论文复现+算法竞赛
Stars: ✭ 341 (-1.16%)
Mutual labels:  jupyter-notebook
Attack Python Client
Python Script to access ATT&CK content available in STIX via a public TAXII server
Stars: ✭ 342 (-0.87%)
Mutual labels:  jupyter-notebook

Website | ArXiv | Get Start

Global-Flow-Local-Attention

The source code for our paper "Deep Image Spatial Transformation for Person Image Generation" (CVPR2020)

We propose a Global-Flow Local-Attention Model for deep image spatial transformation. Our model can be flexibly applied to tasks such as:

  • Pose-Guided Person Image Generation:

Left: generated results of our model; Right: Input source images.

  • Pose-Guided Person Image Animation

Left most: Skeleton Squences. The others: Animation Results.

  • Face Image Animation

Left: Input image; Right: Output results.

  • View Synthesis

Form Left to Right: Input image, Results of Appearance Flow, Results of Ours, Ground-truth images.

News

  • 2020.4.30 Several demos are provided for quick exploration.

  • 2020.4.29 Code for Pose-Guided Person Image Animation is avaliable now!

  • 2020.3.15 We upload the code and trained models of the Face Animation and View Synthesis!

  • 2020.3.3 Project Website and Paper are avaliable!

  • 2020.2.29 Code for PyTorch is available now!

Colab Demo

For a quick exploration of our model, find the online colab demo.

Get Start

1) Installation

Requirements

  • Python 3
  • pytorch (1.0.0)
  • CUDA
  • visdom

Conda installation

# 1. Create a conda virtual environment.
conda create -n gfla python=3.6 -y
source activate gfla

# 2. Install dependency
pip install -r requirement.txt

# 3. Build pytorch Custom CUDA Extensions
./setup.sh

Note: The current code is tested with Tesla V100. If you use a different GPU, you may need to select correct nvcc_args for your GPU when you buil Custom CUDA Extensions. Comment or Uncomment --gencode in block_extractor/setup.py, local_attn_reshape/setup.py, and resample2d_package/setup.py. Please check here for details.

2) Download Resources

We provide the pre-trained weights of our model. The resources are listed as following:

Download the Per-Trained Models and the Demo Images by running the following code:

./download.sh

3) Pose-Guided Person Image Generation

The Pose-Guided Person Image Generation task is to transfer a source person image to a target pose.

Run the demo of this task:

python demo.py \
--name=pose_fashion_checkpoints \
--model=pose \
--attn_layer=2,3 \
--kernel_size=2=5,3=3 \
--gpu_id=0 \
--dataset_mode=fashion \
--dataroot=./dataset/fashion \
--results_dir=./demo_results/fashion

For more training and testing details, please find the PERSON_IMAGE_GENERATION.md

4) Pose-Guided Person Image Animation

The Pose-Guided Person Image Animation task generates a video clip from a still source image according to a driving target sequence. We further model the temporal consistency for this task.

Run the the demo of this task:

python demo.py \
--name=dance_fashion_checkpoints \
--model=dance \
--attn_layer=2,3 \
--kernel_size=2=5,3=3 \
--gpu_id=0 \
--dataset_mode=dance \
--sub_dataset=fashion \
--dataroot=./dataset/danceFashion \
--results_dir=./demo_results/dance_fashion \
--test_list=val_list.csv

For more training and testing details, please find the PERSON_IMAGE_ANIMATION.md.

5) Face Image Animation

Given an input source image and a guidance video sequence depicting the structure movements, our model generating a video containing the specific movements.

Run the the demo of this task:

python demo.py \
--name=face_checkpoints \
--model=face \
--attn_layer=2,3 \
--kernel_size=2=5,3=3 \
--gpu_id=0 \
--dataset_mode=face \
--dataroot=./dataset/FaceForensics \
--results_dir=./demo_results/face 

We use the real video of the FaceForensics dataset. See FACE_IMAGE_ANIMATION.md for more details.

6) Novel View Synthesis

View synthesis requires generating novel views of objects or scenes based on arbitrary input views.

In this task, we use the car and chair categories of the ShapeNet dataset. See VIEW_SYNTHESIS.md for more details.

Citation

@article{ren2020deep,
  title={Deep Image Spatial Transformation for Person Image Generation},
  author={Ren, Yurui and Yu, Xiaoming and Chen, Junming and Li, Thomas H and Li, Ge},
  journal={arXiv preprint arXiv:2003.00696},
  year={2020}
}

@article{ren2020deep,
  title={Deep Spatial Transformation for Pose-Guided Person Image Generation and Animation},
  author={Ren, Yurui and Li, Ge and Liu, Shan and Li, Thomas H},
  journal={IEEE Transactions on Image Processing},
  year={2020},
  publisher={IEEE}
}

Acknowledgement

We build our project base on Vid2Vid. Some dataset preprocessing methods are derived from Pose-Transfer.

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].