All Projects → ZhaoJ9014 → Multi Human Parsing

ZhaoJ9014 / Multi Human Parsing

Licence: mit
🔥🔥Official Repository for Multi-Human-Parsing (MHP)🔥🔥

Programming Languages

javascript
184084 projects - #8 most used programming language

Projects that are alternatives of or similar to Multi Human Parsing

BCNet
Deep Occlusion-Aware Instance Segmentation with Overlapping BiLayers [CVPR 2021]
Stars: ✭ 434 (-14.4%)
Mutual labels:  detection, segmentation, instance-segmentation
Rectlabel Support
RectLabel - An image annotation tool to label images for bounding box object detection and segmentation.
Stars: ✭ 338 (-33.33%)
Mutual labels:  segmentation, annotations, detection
mri-deep-learning-tools
Resurces for MRI images processing and deep learning in 3D
Stars: ✭ 56 (-88.95%)
Mutual labels:  detection, segmentation
Sipmask
SipMask: Spatial Information Preservation for Fast Image and Video Instance Segmentation (ECCV2020)
Stars: ✭ 255 (-49.7%)
Mutual labels:  segmentation, detection
Lidar Bonnetal
Semantic and Instance Segmentation of LiDAR point clouds for autonomous driving
Stars: ✭ 465 (-8.28%)
Mutual labels:  segmentation, semantic
volkscv
A Python toolbox for computer vision research and project
Stars: ✭ 58 (-88.56%)
Mutual labels:  detection, segmentation
Entity
EntitySeg Toolbox: Towards Open-World and High-Quality Image Segmentation
Stars: ✭ 313 (-38.26%)
Mutual labels:  segmentation, instance-segmentation
Holy Edge
Holistically-Nested Edge Detection
Stars: ✭ 277 (-45.36%)
Mutual labels:  segmentation, detection
etiketai
Etiketai is an online tool designed to label images, useful for training AI models
Stars: ✭ 63 (-87.57%)
Mutual labels:  detection, annotations
Erfnet pytorch
Pytorch code for semantic segmentation using ERFNet
Stars: ✭ 304 (-40.04%)
Mutual labels:  segmentation, semantic
Cvpods
All-in-one Toolbox for Computer Vision Research.
Stars: ✭ 277 (-45.36%)
Mutual labels:  segmentation, detection
Awesome Iccv
ICCV2019最新录用情况
Stars: ✭ 305 (-39.84%)
Mutual labels:  segmentation, detection
uoais
Codes of paper "Unseen Object Amodal Instance Segmentation via Hierarchical Occlusion Modeling", ICRA 2022
Stars: ✭ 77 (-84.81%)
Mutual labels:  segmentation, instance-segmentation
mmrazor
OpenMMLab Model Compression Toolbox and Benchmark.
Stars: ✭ 644 (+27.02%)
Mutual labels:  detection, segmentation
unsupervised llamas
Code for https://unsupervised-llamas.com
Stars: ✭ 70 (-86.19%)
Mutual labels:  detection, segmentation
wasr network
WaSR Segmentation Network for Unmanned Surface Vehicles v0.5
Stars: ✭ 32 (-93.69%)
Mutual labels:  semantic, segmentation
Detectron.pytorch
A pytorch implementation of Detectron. Both training from scratch and inferring directly from pretrained Detectron weights are available.
Stars: ✭ 2,805 (+453.25%)
Mutual labels:  segmentation, detection
rgbd person tracking
R-GBD Person Tracking is a ROS framework for detecting and tracking people from a mobile robot.
Stars: ✭ 46 (-90.93%)
Mutual labels:  detection, segmentation
Awesome-Vision-Transformer-Collection
Variants of Vision Transformer and its downstream tasks
Stars: ✭ 124 (-75.54%)
Mutual labels:  detection, segmentation
Fastmaskrcnn
Mask RCNN in TensorFlow
Stars: ✭ 3,069 (+505.33%)
Mutual labels:  segmentation, detection

Multi-Human-Parsing (MHP)

⭐️ ACM MM'18 Best Student Paper

Originality

  • To our best knowledge, we are the first to propose a new Multi-Human Parsing task, corresponding datasets, evaluation metrics and baseline methods.

Task Definition

  • Multi-Human Parsing refers to partitioning a crowd scene image into semantically consistent regions belonging to the body parts or clothes items while differentiating different identities, such that each pixel in the image is assigned a semantic part label, as well as the identity it belongs to. A lot of higher-level applications can be founded upon Multi-Human Parsing, such as virtual reality, automatic production recommendation, video surveillance, and group behavior analysis.

Motivation

  • The Multi-Human Parsing project of Learning and Vision (LV) Group, National University of Singapore (NUS) is proposed to push the frontiers of fine-grained visual understanding of humans in crowd scene.

  • Multi-Human Parsing is significantly different from traditional well-defined object recognition tasks, such as object detection, which only provides coarse-level predictions of object locations (bounding boxes); instance segmentation, which only predicts the instance-level mask without any detailed information on body parts and fashion categories; human parsing, which operates on category-level pixel-wise prediction without differentiating different identities.

  • In real world scenario, the setting of multiple persons with interactions are more realistic and usual. Thus a task, corresponding datasets and baseline methods to consider both the fine-grained semantic information of each individual person and the relationships and interactions of the whole group of people are highly desired.

Multi-Human Parsing (MHP) v1.0 Dataset

  • Statistics: The MHP v1.0 dataset contains 4,980 images, each with at least two persons (average is 3). We randomly choose 980 images and their corresponding annotations as the testing set. The rest form a training set of 3,000 images and a validation set of 1,000 images. For each instance, 18 semantic categories are defined and annotated except for the "background" category, i.e. “hat”, “hair”, “sunglasses”, “upper clothes”, “skirt”, “pants”, “dress”, “belt”, “left shoe”, “right shoe”, “face”, “left leg”, “right leg”, “left arm”, “right arm”, “bag”, “scarf” and “torso skin”. Each instance has a complete set of annotations whenever the corresponding category appears in the current image.

  • WeChat News.

  • Download: The MHP v1.0 dataset is available at google drive and baidu drive (password: cmtp).

  • Please refer to our MHP v1.0 paper (submitted to IJCV) for more details.

Multi-Human Parsing (MHP) v2.0 Dataset

  • Statistics: The MHP v2.0 dataset contains 25,403 images, each with at least two persons (average is 3). We randomly choose 5,000 images and their corresponding annotations as the testing set. The rest form a training set of 15,403 images and a validation set of 5,000 images. For each instance, 58 semantic categories are defined and annotated except for the "background" category, i.e. "cap/hat", "helmet", "face", "hair", "left-arm", "right-arm", "left-hand", "right-hand", "protector", "bikini/bra", "jacket/windbreaker/hoodie", "t-shirt", "polo-shirt", "sweater", "singlet", "torso-skin", "pants", "shorts/swim-shorts", "skirt", "stockings", "socks", "left-boot", "right-boot", "left-shoe", "right-shoe", "left-highheel", "right-highheel", "left-sandal", "right-sandal", "left-leg", "right-leg", "left-foot", "right-foot", "coat", "dress", "robe", "jumpsuit", "other-full-body-clothes", "headwear", "backpack", "ball", "bats", "belt", "bottle", "carrybag", "cases", "sunglasses", "eyewear", "glove", "scarf", "umbrella", "wallet/purse", "watch", "wristband", "tie", "other-accessary", "other-upper-body-clothes" and "other-lower-body-clothes". Each instance has a complete set of annotations whenever the corresponding category appears in the current image. Moreover, 2D human poses with 16 dense key points ("right-shoulder", "right-elbow", "right-wrist", "left-shoulder", "left-elbow", "left-wrist", "right-hip", "right-knee", "right-ankle", "left-hip", "left-knee", "left-ankle", "head", "neck", "spine" and "pelvis". Each key point has a flag indicating whether it is visible-0/occluded-1/out-of-image-2) and head & instance bounding boxes are also provided to facilitate Multi-Human Pose Estimation research.

  • Download: The MHP v2.0 dataset is available at google drive and baidu drive (password: uxrb).

  • Please refer to our MHP v2.0 paper (ACM MM'18 Best Student Paper) for more details.

Evaluation Metrics

  • Multi-Human Parsing: We use two human-centric metrics for multi-human parsing evaluation, which are initially reported by our MHP v1.0 paper. The two metrics are Average Precision based on part (APp) (%) and Percentage of Correctly parsed semantic Parts (PCP) (%). For evaluation code, please refer to the "Evaluation" folder under our "Multi-Human-Parsing_MHP" repository.

  • Multi-Human Pose Estimation: Followed MPII, we use mAP (%) evaluation measure.

CVPR VUHCS2018 Workshop

  • We have organized the CVPR 2018 Workshop on Visual Understanding of Humans in Crowd Scene (VUHCS 2018). This workshop is collaborated by NUS, CMU and SYSU. Based on VUHCS 2017, we have further strengthened this Workshop by augmenting it with 5 competition tracks: the single-person human parsing, the multi-person human parsing, the single-person pose estimation, the multi-human pose estimation and the fine-grained multi-human parsing.

  • Result Submission & Leaderboard.

  • WeChat News.

Citation

  • Please consult and consider citing the following papers:

    @article{zhao2018understanding,
    title={Understanding Humans in Crowded Scenes: Deep Nested Adversarial Learning and A New Benchmark for Multi-Human Parsing},
    author={Zhao, Jian and Li, Jianshu and Cheng, Yu and Zhou, Li and Sim, Terence and Yan, Shuicheng and Feng, Jiashi},
    journal={arXiv preprint arXiv:1804.03287},
    year={2018}
    }
    
    
    @article{li2017towards,
    title={Multi-Human Parsing in the Wild},
    author={Li, Jianshu and Zhao, Jian and Wei, Yunchao and Lang, Congyan and Li, Yidong and Sim, Terence and Yan, Shuicheng and Feng, Jiashi},
    journal={arXiv preprint arXiv:1705.07206},
    year={2017}
    }
    
Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].