All Projects → wywu → Lab

wywu / Lab

Licence: other
[CVPR 2018] Look at Boundary: A Boundary-Aware Face Alignment Algorithm

Projects that are alternatives of or similar to Lab

Facekit
Implementations of PCN (an accurate real-time rotation-invariant face detector) and other face-related algorithms
Stars: ✭ 1,028 (+7.53%)
Mutual labels:  cvpr2018, face-alignment
Taadpapers
Must-read Papers on Textual Adversarial Attack and Defense
Stars: ✭ 800 (-16.32%)
Mutual labels:  adversarial-learning
Faceswap
3D face swapping implemented in Python
Stars: ✭ 508 (-46.86%)
Mutual labels:  face-alignment
Gibsonenv
Gibson Environments: Real-World Perception for Embodied Agents
Stars: ✭ 666 (-30.33%)
Mutual labels:  cvpr2018
Lemniscate.pytorch
Unsupervised Feature Learning via Non-parametric Instance Discrimination
Stars: ✭ 532 (-44.35%)
Mutual labels:  cvpr2018
Polyrnn Pp
Inference Code for Polygon-RNN++ (CVPR 2018)
Stars: ✭ 704 (-26.36%)
Mutual labels:  cvpr2018
Deepalignmentnetwork
A deep neural network for face alignment
Stars: ✭ 480 (-49.79%)
Mutual labels:  face-alignment
Vc
Code for CVPR'18 "Grounding Referring Expressions in Images by Variational Context"
Stars: ✭ 26 (-97.28%)
Mutual labels:  cvpr2018
Neural Structured Learning
Training neural models with structured signals.
Stars: ✭ 790 (-17.36%)
Mutual labels:  adversarial-learning
Adaptsegnet
Learning to Adapt Structured Output Space for Semantic Segmentation, CVPR 2018 (spotlight)
Stars: ✭ 654 (-31.59%)
Mutual labels:  adversarial-learning
Hrnet Facial Landmark Detection
This is an official implementation of facial landmark detection for our TPAMI paper "Deep High-Resolution Representation Learning for Visual Recognition". https://arxiv.org/abs/1908.07919
Stars: ✭ 647 (-32.32%)
Mutual labels:  face-alignment
Face Alignment
🔥 2D and 3D Face alignment library build using pytorch
Stars: ✭ 5,417 (+466.63%)
Mutual labels:  face-alignment
Neurec
Next RecSys Library
Stars: ✭ 731 (-23.54%)
Mutual labels:  adversarial-learning
Fots.pytorch
FOTS Pytorch Implementation
Stars: ✭ 513 (-46.34%)
Mutual labels:  cvpr2018
Advertorch
A Toolbox for Adversarial Robustness Research
Stars: ✭ 826 (-13.6%)
Mutual labels:  adversarial-learning
Awesome Cvpr Paper
CVPR 论文收集,包含但不限于2021、2020、2019、2018、2017文章
Stars: ✭ 493 (-48.43%)
Mutual labels:  cvpr2018
Tf Dann
Domain-Adversarial Neural Network in Tensorflow
Stars: ✭ 556 (-41.84%)
Mutual labels:  adversarial-learning
Polyrnn Pp Pytorch
PyTorch training/tool code for Polygon-RNN++ (CVPR 2018)
Stars: ✭ 672 (-29.71%)
Mutual labels:  cvpr2018
Advanced Gradient Obfuscating
Take further steps in the arms race of adversarial examples with only preprocessing.
Stars: ✭ 28 (-97.07%)
Mutual labels:  adversarial-learning
Ffnet
Implementation of our CVPR 2018 Paper (FFNet: Video Fast-Forwarding via Reinforcement Learning)
Stars: ✭ 17 (-98.22%)
Mutual labels:  cvpr2018

Look at Boundary: A Boundary-Aware Face Alignment Algorithm

Created by Wayne Wu at Tsinghua University.

We present a novel boundary-aware face alignment algorithm by utilising boundary lines as the geometric structure of a human face to help facial landmark localisation. Unlike the conventional heatmap based method and regression based method, our approach derives face landmarks from boundary lines which remove the ambiguities in the landmark definition. Three questions are explored and answered by this work: 1. Why use boundary? 2. How do you use boundary? 3. What is the relationship between boundary estimation and landmarks localisation?

Our boundary-aware face alignment algorithm achieves 3.49% mean error on 300-W Fullset, which outperforms state-of-the-art methods by a large margin. Our method can also easily integrate information from other datasets. By utilising boundary information of 300-W dataset, our method achieves 3.92% mean error with 0.39% failure rate on COFW dataset, and 1.25% mean error on AFLW-Full dataset. Moreover, we propose a new dataset Wider Facial Landmark in the Wild (WFLW) to unify training and testing across different factors, including poses, expressions, illuminations, makeups, occlusions, and blurriness. A detailed description of the system can be found in our paper.

Citation

If you use this code or WFLW dataset for your research, please cite our papers.

@inproceedings{wayne2018lab,
 author = {Wu, Wayne and Qian, Chen and Yang, Shuo and Wang, Quan and Cai, Yici and Zhou, Qiang},
 title = {Look at Boundary: A Boundary-Aware Face Alignment Algorithm},
 booktitle = {CVPR},
 month = June,
 year = {2018}
}

Prerequisites

  • Linux
  • Python 2 or 3
  • CPU or NVIDIA GPU + CUDA CuDNN

Getting Started

Installing

  1. Install prerequisites for Caffe (http://caffe.berkeleyvision.org/installation.html#prequequisites)
  2. Modified-caffe for LAB (https://github.com/wywu/LAB.git)

Wider Facial Landmark in the Wild (WFLW) Dataset Download

Wider Facial Landmarks in-the-wild (WFLW) is a new proposed face dataset. It contains 10000 faces (7500 for training and 2500 for testing) with 98 fully manual annotated landmarks.

  1. WFLW Training and Testing images [Google Drive] [Baidu Drive]
  2. WFLW Face Annotations
  3. Unzip above two packages and put them on './datasets/WFLW/'

Simply run this script to download annotations of WFLW

#! ./scripts/download/download_wflw_annotation.sh
bash ./scripts/download/download_wflw_annotation.sh WFLW

Testing LAB on WFLW

We supply two pretrained models:

WFLW_final: The final model evaluated on WFLW in the paper.
WFLW_wo_mp: The simplified model without Message Passing layer which is much easier to read.
  1. Download pretrained models: WFLW_final and WFLW_wo_mp.
#! ./scripts/download/download_wflw_model.sh
bash ./scripts/download/download_wflw_model.sh WFLW
  1. Test the model:
#! ./scripts/test/run_test_on_wflw.sh
bash ./scripts/test/run_test_on_wflw.sh final
bash ./scripts/test/run_test_on_wflw.sh wo_mp

The testing results will be saved to text files here: ./evaluation/WFLW/.

Results of WFLW_final model on WFLW

Training

For company security considerations, it is with regret that we are not allowed to release the training scripts. However, for training, all you need to do is using this released code and adding data augmentation described in the paper. Taking the released prototxt file as reference and using the hyper-parameters described in the paper, we think it is easy to reproduce the reported performance. If you have any question about the training process, please feel free to contact us.

To Do List

Supported dataset

Supported models

Questions

Please contact [email protected]

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].