All Projects → HRNet → Hrnet Facial Landmark Detection

HRNet / Hrnet Facial Landmark Detection

Licence: mit
This is an official implementation of facial landmark detection for our TPAMI paper "Deep High-Resolution Representation Learning for Visual Recognition". https://arxiv.org/abs/1908.07919

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to Hrnet Facial Landmark Detection

ECT-FaceAlignment
[TIFS 2018] Combining Data-driven and Model-driven Methods for Robust Facial Landmark Detection
Stars: ✭ 28 (-95.67%)
Mutual labels:  face-alignment
Flame Fitting
Example code for the FLAME 3D head model. The code demonstrates how to sample 3D heads from the model, fit the model to 3D keypoints and 3D scans.
Stars: ✭ 269 (-58.42%)
Mutual labels:  face-alignment
Pytorch face landmark
Fast and accurate face landmark detection library using PyTorch; Support 68-point semi-frontal and 39-point profile landmark detection; Support both coordinate-based and heatmap-based inference; Up to 100 FPS landmark inference speed with SOTA face detector on CPU.
Stars: ✭ 388 (-40.03%)
Mutual labels:  face-alignment
DeepVTB
🌌 OpenVTuber-虚拟アイドル共享计划 An application of real-time face and gaze analyzation via deep nerual networks.
Stars: ✭ 32 (-95.05%)
Mutual labels:  face-alignment
retinaface
RetinaFace: Deep Face Detection Library for Python
Stars: ✭ 242 (-62.6%)
Mutual labels:  face-alignment
3ddfa
The PyTorch improved version of TPAMI 2017 paper: Face Alignment in Full Pose Range: A 3D Total Solution.
Stars: ✭ 3,154 (+387.48%)
Mutual labels:  face-alignment
Face
I have published my face related codes in this repository
Stars: ✭ 53 (-91.81%)
Mutual labels:  face-alignment
Faceswap
3D face swapping implemented in Python
Stars: ✭ 508 (-21.48%)
Mutual labels:  face-alignment
iqiyi-vid-challenge
Code for IQIYI-VID(IQIYI Video Person Identification) Challenge Implemented in Python and MXNet
Stars: ✭ 45 (-93.04%)
Mutual labels:  face-alignment
Openvtuber
虚拟爱抖露(アイドル)共享计划, 是基于单目RGB摄像头的人眼与人脸特征点检测算法, 在实时3D面部捕捉以及模型驱动领域的应用.
Stars: ✭ 365 (-43.59%)
Mutual labels:  face-alignment
enhanced-ssh-mxnet
The MXNet Implementation of Enhanced SSH (ESSH) for Face Detection and Alignment
Stars: ✭ 54 (-91.65%)
Mutual labels:  face-alignment
Laser-Eye
Gaze Estimation via Deep Neural Networks
Stars: ✭ 79 (-87.79%)
Mutual labels:  face-alignment
High Performance Face Recognition
🔥🔥Several High-Performance Models for Unconstrained/Large-Scale/Low-Shot Face Recognition🔥🔥
Stars: ✭ 309 (-52.24%)
Mutual labels:  face-alignment
Face-alignment-Trees
This is the C++ implement of the paper: Face Detection, Pose Estimation, and Landmark Localization in the Wild
Stars: ✭ 17 (-97.37%)
Mutual labels:  face-alignment
Face Pose Net
Estimate 3D face pose (6DoF) or 11 parameters of 3x4 projection matrix by a Convolutional Neural Network
Stars: ✭ 464 (-28.28%)
Mutual labels:  face-alignment
SupervisedDescentMethod
No description or website provided.
Stars: ✭ 19 (-97.06%)
Mutual labels:  face-alignment
Dest
🐼 One Millisecond Deformable Shape Tracking Library (DEST)
Stars: ✭ 276 (-57.34%)
Mutual labels:  face-alignment
Face Alignment
🔥 2D and 3D Face alignment library build using pytorch
Stars: ✭ 5,417 (+737.25%)
Mutual labels:  face-alignment
Deepalignmentnetwork
A deep neural network for face alignment
Stars: ✭ 480 (-25.81%)
Mutual labels:  face-alignment
Peppa pig face engine
A simple face detect and alignment method, which is easy and stable.
Stars: ✭ 329 (-49.15%)
Mutual labels:  face-alignment

High-resolution networks (HRNets) for facial landmark detection

News

Introduction

This is the official code of High-Resolution Representations for Facial Landmark Detection. We extend the high-resolution representation (HRNet) [1] by augmenting the high-resolution representation by aggregating the (upsampled) representations from all the parallel convolutions, leading to stronger representations. The output representations are fed into classifier. We evaluate our methods on four datasets, COFW, AFLW, WFLW and 300W.

Performance

ImageNet pretrained models

HRNetV2 ImageNet pretrained models are now available! Codes and pretrained models are in HRNets for Image Classification

We adopt HRNetV2-W18(#Params=9.3M, GFLOPs=4.3G) for facial landmark detection on COFW, AFLW, WFLW and 300W.

COFW

The model is trained on COFW train and evaluated on COFW test.

Model NME FR0.1 pretrained model model
HRNetV2-W18 3.45 0.20 HRNetV2-W18 HR18-COFW.pth

AFLW

The model is trained on AFLW train and evaluated on AFLW full and frontal.

Model NMEfull NMEfrontal pretrained model model
HRNetV2-W18 1.57 1.46 HRNetV2-W18 HR18-AFLW.pth

WFLW

NME test pose illumination occlution blur makeup expression pretrained model model
HRNetV2-W18 4.60 7.86 4.57 5.42 5.36 4.26 4.78 HRNetV2-W18 HR18-WFLW.pth

300W

NME common challenge full test pretrained model model
HRNetV2-W18 2.91 5.11 3.34 3.85 HRNetV2-W18 HR18-300W.pth

Quick start

Environment

This code is developed using on Python 3.6 and PyTorch 1.0.0 on Ubuntu 16.04 with NVIDIA GPUs. Training and testing are performed using 1 NVIDIA P40 GPU with CUDA 9.0 and cuDNN 7.0. Other platforms or GPUs are not fully tested.

Install

  1. Install PyTorch 1.0 following the official instructions
  2. Install dependencies
pip install -r requirements.txt
  1. Clone the project
git clone https://github.com/HRNet/HRNet-Facial-Landmark-Detection.git

HRNetV2 pretrained models

cd HRNet-Facial-Landmark-Detection
# Download pretrained models into this folder
mkdir hrnetv2_pretrained

Data

  1. You need to download the annotations files which have been processed from OneDrive, Cloudstor, and BaiduYun(Acess Code:ypxg).

  2. You need to download images (300W, AFLW, WFLW) from official websites and then put them into images folder for each dataset.

Your data directory should look like this:

HRNet-Facial-Landmark-Detection
-- lib
-- experiments
-- tools
-- data
   |-- 300w
   |   |-- face_landmarks_300w_test.csv
   |   |-- face_landmarks_300w_train.csv
   |   |-- face_landmarks_300w_valid.csv
   |   |-- face_landmarks_300w_valid_challenge.csv
   |   |-- face_landmarks_300w_valid_common.csv
   |   |-- images
   |-- aflw
   |   |-- face_landmarks_aflw_test.csv
   |   |-- face_landmarks_aflw_test_frontal.csv
   |   |-- face_landmarks_aflw_train.csv
   |   |-- images
   |-- cofw
   |   |-- COFW_test_color.mat
   |   |-- COFW_train_color.mat  
   |-- wflw
   |   |-- face_landmarks_wflw_test.csv
   |   |-- face_landmarks_wflw_test_blur.csv
   |   |-- face_landmarks_wflw_test_expression.csv
   |   |-- face_landmarks_wflw_test_illumination.csv
   |   |-- face_landmarks_wflw_test_largepose.csv
   |   |-- face_landmarks_wflw_test_makeup.csv
   |   |-- face_landmarks_wflw_test_occlusion.csv
   |   |-- face_landmarks_wflw_train.csv
   |   |-- images

Train

Please specify the configuration file in experiments (learning rate should be adjusted when the number of GPUs is changed).

python tools/train.py --cfg <CONFIG-FILE>
# example:
python tools/train.py --cfg experiments/wflw/face_alignment_wflw_hrnet_w18.yaml

Test

python tools/test.py --cfg <CONFIG-FILE> --model-file <MODEL WEIGHT> 
# example:
python tools/test.py --cfg experiments/wflw/face_alignment_wflw_hrnet_w18.yaml --model-file HR18-WFLW.pth

Other applications of HRNets (codes and models):

Citation

If you find this work or code is helpful in your research, please cite:

@inproceedings{SunXLW19,
  title={Deep High-Resolution Representation Learning for Human Pose Estimation},
  author={Ke Sun and Bin Xiao and Dong Liu and Jingdong Wang},
  booktitle={CVPR},
  year={2019}
}

@article{WangSCJDZLMTWLX19,
  title={Deep High-Resolution Representation Learning for Visual Recognition},
  author={Jingdong Wang and Ke Sun and Tianheng Cheng and 
          Borui Jiang and Chaorui Deng and Yang Zhao and Dong Liu and Yadong Mu and 
          Mingkui Tan and Xinggang Wang and Wenyu Liu and Bin Xiao},
  journal   = {TPAMI}
  year={2019}
}

Reference

[1] Deep High-Resolution Representation Learning for Visual Recognition. Jingdong Wang, Ke Sun, Tianheng Cheng, Borui Jiang, Chaorui Deng, Yang Zhao, Dong Liu, Yadong Mu, Mingkui Tan, Xinggang Wang, Wenyu Liu, Bin Xiao. Accepted by TPAMI. download

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].