All Projects → SysCV → pcan

SysCV / pcan

Licence: Apache-2.0 license
Prototypical Cross-Attention Networks for Multiple Object Tracking and Segmentation, NeurIPS 2021 Spotlight

Programming Languages

python
139335 projects - #7 most used programming language
shell
77523 projects

Projects that are alternatives of or similar to pcan

UniTrack
[NeurIPS'21] Unified tracking framework with a single appearance model. It supports Single Object Tracking (SOT), Video Object Segmentation (VOS), Multi-Object Tracking (MOT), Multi-Object Tracking and Segmentation (MOTS), Pose Tracking, Video Instance Segmentation (VIS), and class-agnostic MOT (e.g. TAO dataset).
Stars: ✭ 293 (-0.34%)
Mutual labels:  video-instance-segmentation, multi-object-tracking-segmentation
Sipmask
SipMask: Spatial Information Preservation for Fast Image and Video Instance Segmentation (ECCV2020)
Stars: ✭ 255 (-13.27%)
Mutual labels:  tracking, segmentation
pedx
Python tools for working with PedX dataset.
Stars: ✭ 26 (-91.16%)
Mutual labels:  tracking, segmentation
Maskfusion
MaskFusion: Real-Time Recognition, Tracking and Reconstruction of Multiple Moving Objects
Stars: ✭ 404 (+37.41%)
Mutual labels:  tracking, segmentation
Trackit
[ECCV'20] Ocean: Object-aware Anchor-Free Tracking
Stars: ✭ 424 (+44.22%)
Mutual labels:  tracking, segmentation
Dipy
DIPY is the paragon 3D/4D+ imaging library in Python. Contains generic methods for spatial normalization, signal processing, machine learning, statistical analysis and visualization of medical images. Additionally, it contains specialized methods for computational anatomy including diffusion, perfusion and structural imaging.
Stars: ✭ 417 (+41.84%)
Mutual labels:  tracking, segmentation
Co Fusion
Co-Fusion: Real-time Segmentation, Tracking and Fusion of Multiple Objects
Stars: ✭ 400 (+36.05%)
Mutual labels:  tracking, segmentation
Deep Learning For Tracking And Detection
Collection of papers, datasets, code and other resources for object tracking and detection using deep learning
Stars: ✭ 1,920 (+553.06%)
Mutual labels:  tracking, segmentation
Awesome-Vision-Transformer-Collection
Variants of Vision Transformer and its downstream tasks
Stars: ✭ 124 (-57.82%)
Mutual labels:  tracking, segmentation
farm-animal-tracking
Farm Animal Tracking (FAT)
Stars: ✭ 19 (-93.54%)
Mutual labels:  tracking
Kalman.jl
Flexible filtering and smoothing in Julia
Stars: ✭ 62 (-78.91%)
Mutual labels:  tracking
Better-Fox
An up-to-date user.js to speed up and secure Firefox
Stars: ✭ 158 (-46.26%)
Mutual labels:  tracking
DeepSegmentor
Sequence Segmentation using Joint RNN and Structured Prediction Models (ICASSP 2017)
Stars: ✭ 17 (-94.22%)
Mutual labels:  segmentation
traceml
Engine for ML/Data tracking, visualization, dashboards, and model UI for Polyaxon.
Stars: ✭ 445 (+51.36%)
Mutual labels:  tracking
nicMSlesions
Easy multiple sclerosis white matter lesion segmentation using convolutional deep neural networks.
Stars: ✭ 33 (-88.78%)
Mutual labels:  segmentation
MiVOS
[CVPR 2021] Modular Interactive Video Object Segmentation: Interaction-to-Mask, Propagation and Difference-Aware Fusion. Semi-supervised VOS as well!
Stars: ✭ 302 (+2.72%)
Mutual labels:  segmentation
Basic-Image-Processing
Implementation of Basic Digital Image Processing Tasks in Python / OpenCV
Stars: ✭ 102 (-65.31%)
Mutual labels:  segmentation
Polygonization-by-Frame-Field-Learning
This repository contains the code for our fast polygonal building extraction from overhead images pipeline.
Stars: ✭ 161 (-45.24%)
Mutual labels:  segmentation
color-pop
🌈 Automatic Color Pop effect on any image inspired by Google Photos
Stars: ✭ 21 (-92.86%)
Mutual labels:  segmentation
hair-dye
Neural Network for Dying Hair💈
Stars: ✭ 45 (-84.69%)
Mutual labels:  segmentation

PWC

Prototypical Cross-Attention Networks for Multiple Object Tracking and Segmentation

This is the offical implementation of paper PCAN for MOTS. PCAN is also served as the baseline method in BDD100K tracking challenges at CVPR 2022.

We also present a trailer that consists of method illustrations and tracking & segmentation visualizations. Our project website contains more information: vis.xyz/pub/pcan.

Prototypical Cross-Attention Networks for Multiple Object Tracking and Segmentation
NeurIPS 2021, Spotlight
Lei Ke, Xia Li, Martin Danelljan, Yu-Wing Tai, Chi-Keung Tang, Fisher Yu

Abstract

Multiple object tracking and segmentation requires detecting, tracking, and segmenting objects belonging to a set of given classes. Most approaches only exploit the temporal dimension to address the association problem, while relying on single frame predictions for the segmentation mask itself. We propose Prototypical Cross-Attention Network (PCAN), capable of leveraging rich spatio-temporal information for online multiple object tracking and segmentation. PCAN first distills a space-time memory into a set of prototypes and then employs cross-attention to retrieve rich information from the past frames. To segment each object, PCAN adopts a prototypical appearance module to learn a set of contrastive foreground and background prototypes, which are then propagated over time. Extensive experiments demonstrate that PCAN outperforms current video instance tracking and segmentation competition winners on both Youtube-VIS and BDD100K datasets, and shows efficacy to both one-stage and two-stage segmentation frameworks.

Prototypical Cross-Attention Networks (PCAN)

Main results

Results on BDD100K Benchmark

Detector mMOTSA-val mIDF1-val ID Sw.-val Scores-val mMOTSA-test mIDF1-test ID Sw.-test Scores-test Config Weights Preds Visuals
ResNet-50 28.1 45.4 874 scores 31.9 50.4 845 scores config model | MD5 preds visuals

Installation

Please refer to INSTALL.md for installation instructions.

Usages

Please refer to GET_STARTED.md for dataset preparation and detailed running (training, testing, visualization, etc.) instructions.

Related links

Youtube Video | Bilibili Video| Zhihu Reading

Citation

If you find PCAN useful in your research or refer to the provided baseline results, please star this repository and consider citing 📝:

@inproceedings{pcan,
  title={Prototypical Cross-Attention Networks for Multiple Object Tracking and Segmentation},
  author={Ke, Lei and Li, Xia and Danelljan, Martin and Tai, Yu-Wing and Tang, Chi-Keung and Yu, Fisher},
  booktitle={Advances in Neural Information Processing Systems},
  year={2021}
}
Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].