All Projects → MVIG-SJTU → Rmpe

MVIG-SJTU / Rmpe

Licence: other
RMPE: Regional Multi-person Pose Estimation, forked from Caffe. Research purpose only.

Projects that are alternatives of or similar to Rmpe

FashionAI-Keypoint
fashionAI clothes keypoint detection
Stars: ✭ 19 (-92.64%)
Mutual labels:  pose-estimation
openpifpaf
Official implementation of "OpenPifPaf: Composite Fields for Semantic Keypoint Detection and Spatio-Temporal Association" in PyTorch.
Stars: ✭ 900 (+248.84%)
Mutual labels:  pose-estimation
HyperFace-TensorFlow-implementation
HyperFace
Stars: ✭ 68 (-73.64%)
Mutual labels:  pose-estimation
OPVO
Sample code of BMVC 2017 paper: "Visual Odometry with Drift-Free Rotation Estimation Using Indoor Scene Regularities"
Stars: ✭ 40 (-84.5%)
Mutual labels:  pose-estimation
Indoor-SfMLearner
[ECCV'20] Patch-match and Plane-regularization for Unsupervised Indoor Depth Estimation
Stars: ✭ 115 (-55.43%)
Mutual labels:  pose-estimation
tensorrt-examples
TensorRT Examples (TensorRT, Jetson Nano, Python, C++)
Stars: ✭ 31 (-87.98%)
Mutual labels:  pose-estimation
LEMO
Official Pytorch implementation for 2021 ICCV paper "Learning Motion Priors for 4D Human Body Capture in 3D Scenes" and trained models / data
Stars: ✭ 149 (-42.25%)
Mutual labels:  pose-estimation
binPicking 3dseg
separate from 6dpose repo for deployment
Stars: ✭ 19 (-92.64%)
Mutual labels:  pose-estimation
EgoNet
Official project website for the CVPR 2021 paper "Exploring intermediate representation for monocular vehicle pose estimation"
Stars: ✭ 111 (-56.98%)
Mutual labels:  pose-estimation
rmpe dataset server
Realtime Multi-Person Pose Estimation data server. Used as a training and validation data provider in training process.
Stars: ✭ 14 (-94.57%)
Mutual labels:  pose-estimation
MobileNetV2-PoseEstimation
Tensorflow based Fast Pose estimation. OpenVINO, Tensorflow Lite, NCS, NCS2 + Python.
Stars: ✭ 99 (-61.63%)
Mutual labels:  pose-estimation
Barracuda-PoseNet
PoseNet Using Unity MLAgents Barracuda Engine
Stars: ✭ 31 (-87.98%)
Mutual labels:  pose-estimation
StarMap
StarMap for Category-Agnostic Keypoint and Viewpoint Estimation
Stars: ✭ 97 (-62.4%)
Mutual labels:  pose-estimation
Primer-MotionCapture
A Primer on Motion Capture with Deep Learning:Principles, Pitfalls and Perspectives
Stars: ✭ 19 (-92.64%)
Mutual labels:  pose-estimation
deep underwater localization
Source Code for "DeepURL: Deep Pose Estimation Framework for Underwater Relative Localization", submitted to IROS 2020
Stars: ✭ 13 (-94.96%)
Mutual labels:  pose-estimation
sc depth pl
Pytorch Lightning Implementation of SC-Depth (V1, V2...) for Unsupervised Monocular Depth Estimation.
Stars: ✭ 86 (-66.67%)
Mutual labels:  pose-estimation
ViPNAS
The official repo for CVPR2021——ViPNAS: Efficient Video Pose Estimation via Neural Architecture Search.
Stars: ✭ 32 (-87.6%)
Mutual labels:  pose-estimation
EgoPose
Official PyTorch Implementation of "Ego-Pose Estimation and Forecasting as Real-Time PD Control". ICCV 2019.
Stars: ✭ 65 (-74.81%)
Mutual labels:  pose-estimation
handobjectconsist
[cvpr 20] Demo, training and evaluation code for joint hand-object pose estimation in sparsely annotated videos
Stars: ✭ 100 (-61.24%)
Mutual labels:  pose-estimation
ONNX-Mobile-Human-Pose-3D
Python scripts for performing 3D human pose estimation using the Mobile Human Pose model in ONNX.
Stars: ✭ 69 (-73.26%)
Mutual labels:  pose-estimation

RMPE: Regional Multi-person Pose Estimation

By Hao-Shu Fang, Shuqin Xie, Yu-Wing Tai, Cewu Lu.

New version AlphaPose is released. The accuracy is 10 mAP higher than this repo. Please move to https://github.com/MVIG-SJTU/AlphaPose

RMPE is a two steps framework for the task of multi-person pose estimation. You can use the code to train/evaluate a model for pose estimation task. For more details, please refer to our arxiv paper.

RMPE Framework

Results

Video results available here

Results on MPII dataset:

Method MPII full test mAP s/frame
Iqbal & Gall, ECCVw'16 43.1 10
DeeperCut, ECCV16 59.5 485
RMPE 76.7 1.5
Results on COCO test-dev 2015:
Method AP @0.5:0.95 AP @0.5 AP @0.75
RMPE 61.8 83.7 69.8

Contents

  1. Installation
  2. Preparation
  3. Demo
  4. Train/Eval
  5. Citation
  6. Acknowledgements

Installation

  1. Get the code. We will call the directory that you cloned Caffe into $CAFFE_ROOT
git clone https://github.com/MVIG-SJTU/RMPE.git
cd RMPE

you can also download code from

git clone https://github.com/Fang-Haoshu/RMPE.git
  1. Build the code. Please follow Caffe instruction to install all necessary packages and build it.
# Modify Makefile.config according to your Caffe installation.
# Note that the SSTN module currently ONLY have GPU implementation so you need to make&run it with GPU
cp Makefile.config.example Makefile.config
make -j8
# Make sure to include $CAFFE_ROOT/python to your PYTHONPATH.
make py
make test -j8
make runtest -j8
# If you have multiple GPUs installed in your machine, make runtest might fail. If so, try following:
export CUDA_VISIBLE_DEVICES=0; make runtest -j8
# If you have error: "Check failed: error == cudaSuccess (10 vs. 0)  invalid device ordinal",
# first make sure you have the specified GPUs, or try following if you have multiple GPUs:
unset CUDA_VISIBLE_DEVICES

Preparation

For demo only

  1. Download pre-trained human detector(Google drive|Baidu cloud) and SPPE+SSTN caffe model(Google drive|Baidu cloud). By default, we assume the models are stored in $CAFFE_ROOT/models/VGG_SSD/ and $CAFFE_ROOT/models/SPPE/ accordingly.

For train/eval

This part of our model is implemented in Torch7. Please refer to this repo for more details.

Demo

Our experiments use both Caffe and Torch7. But we implement the whole framework in Caffe so you can run the demo easily. Note: The current caffe model of SPPE use the 2-stacked hourglass network which has a lower precision. We will be grateful if anyone can help to transfer new torch model to caffe.

  1. Run the ipython notebook. It will show you how our whole framework works
cd $CAFFE_ROOT
# it shows how our framework works
jupyter notebook examples/rmpe/Regional\ Multi-person\ Pose\ Estimation.ipynb
  1. Run the python program for more results
python examples/rmpe/demo.py

Train/Eval

  1. Train SPPE+SSTN. This part of our model is implemented in Torch7. Please refer to this repo for more details. We will call the directory that you cloned the repo into $SPPE_ROOT. I have written an implementation in Caffe. You can email me for the script.

  2. Evaluate the model. You can modify line 45 in demo.py to evaluate our framework on whole test set. But the results will be different. To reproduce our results reported in our paper:

# First get the result of human detector
cd $CAFFE_ROOT
jupyter notebook examples/rmpe/human_detection.ipynb
# Then move the results to $SPPE_ROOT/predict/annot/
mv examples/rmpe/mpii-test0.09 $SPPE_ROOT/predict/annot/
# Next, do single person human estimation
cd $SPPE_ROOT/predict
th main.lua predict-test
#Finally, do pose NMS
python batch_nms.py
#our result is stored in txt format, to evaluate, Download MPII toolkit and put it in current directory
matlab
#In matlab
setpred()

Citation

Please cite the paper in your publications if it helps your research:

@inproceedings{fang2017rmpe,
  title={{RMPE}: Regional Multi-person Pose Estimation},
  author={Fang, Hao-Shu and Xie, Shuqin and Tai, Yu-Wing and Lu, Cewu},
  booktitle={ICCV},
  year={2017}
}

Acknowledgements

Thanks to Wei Liu, Alejandro Newell, Pfister, T., Kaichun Mo, Maxime Oquab for contributing their codes. Thanks to the authors of Caffe and Torch7!

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].