All Projects → JunweiLiang → Multiverse

JunweiLiang / Multiverse

Licence: apache-2.0
Dataset, code and model for the CVPR'20 paper "The Garden of Forking Paths: Towards Multi-Future Trajectory Prediction". And for the ECCV'20 SimAug paper.

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to Multiverse

just-ask
[TPAMI Special Issue on ICCV 2021 Best Papers, Oral] Just Ask: Learning to Answer Questions from Millions of Narrated Videos
Stars: ✭ 57 (-56.49%)
Mutual labels:  video-understanding
Activity Recognition With Cnn And Rnn
Temporal Segments LSTM and Temporal-Inception for Activity Recognition
Stars: ✭ 415 (+216.79%)
Mutual labels:  video-understanding
Temporal Shift Module
[ICCV 2019] TSM: Temporal Shift Module for Efficient Video Understanding
Stars: ✭ 1,282 (+878.63%)
Mutual labels:  video-understanding
CP-360-Weakly-Supervised-Saliency
CP-360-Weakly-Supervised-Saliency
Stars: ✭ 20 (-84.73%)
Mutual labels:  video-understanding
Awesome Action Recognition
A curated list of action recognition and related area resources
Stars: ✭ 3,202 (+2344.27%)
Mutual labels:  video-understanding
Mmaction2
OpenMMLab's Next Generation Video Understanding Toolbox and Benchmark
Stars: ✭ 684 (+422.14%)
Mutual labels:  video-understanding
MTL-AQA
What and How Well You Performed? A Multitask Learning Approach to Action Quality Assessment [CVPR 2019]
Stars: ✭ 38 (-70.99%)
Mutual labels:  video-understanding
I3d finetune
TensorFlow code for finetuning I3D model on UCF101.
Stars: ✭ 128 (-2.29%)
Mutual labels:  video-understanding
Video Understanding Dataset
A collection of recent video understanding datasets, under construction!
Stars: ✭ 387 (+195.42%)
Mutual labels:  video-understanding
Temporally Language Grounding
A Pytorch implemention for some state-of-the-art models for" Temporally Language Grounding in Untrimmed Videos"
Stars: ✭ 73 (-44.27%)
Mutual labels:  video-understanding
DIN-Group-Activity-Recognition-Benchmark
A new codebase for Group Activity Recognition. It contains codes for ICCV 2021 paper: Spatio-Temporal Dynamic Inference Network for Group Activity Recognition and some other methods.
Stars: ✭ 26 (-80.15%)
Mutual labels:  video-understanding
DEAR
[ICCV 2021 Oral] Deep Evidential Action Recognition
Stars: ✭ 36 (-72.52%)
Mutual labels:  video-understanding
Tsn Pytorch
Temporal Segment Networks (TSN) in PyTorch
Stars: ✭ 895 (+583.21%)
Mutual labels:  video-understanding
Awesome-Temporally-Language-Grounding
A curated list of “Temporally Language Grounding” and related area
Stars: ✭ 97 (-25.95%)
Mutual labels:  video-understanding
Temporal Segment Networks
Code & Models for Temporal Segment Networks (TSN) in ECCV 2016
Stars: ✭ 1,287 (+882.44%)
Mutual labels:  video-understanding
STCNet
STCNet: Spatio-Temporal Cross Network for Industrial Smoke Detection
Stars: ✭ 29 (-77.86%)
Mutual labels:  video-understanding
Action Detection
temporal action detection with SSN
Stars: ✭ 597 (+355.73%)
Mutual labels:  video-understanding
Mmaction
An open-source toolbox for action understanding based on PyTorch
Stars: ✭ 1,711 (+1206.11%)
Mutual labels:  video-understanding
Movienet Tools
Tools for movie and video research
Stars: ✭ 113 (-13.74%)
Mutual labels:  video-understanding
Tdn
[CVPR 2021] TDN: Temporal Difference Networks for Efficient Action Recognition
Stars: ✭ 72 (-45.04%)
Mutual labels:  video-understanding

Multiverse

This repository contains the code and models for the following CVPR'20 paper:

The Garden of Forking Paths: Towards Multi-Future Trajectory Prediction
Junwei Liang, Lu Jiang, Kevin Murphy, Ting Yu, Alexander Hauptmann

You can find more information at our Project Page and the blog.

The SimAug (ECCV'20) project is here.

If you find this code useful in your research then please cite

@inproceedings{liang2020garden,
  title={The Garden of Forking Paths: Towards Multi-Future Trajectory Prediction},
  author={Liang, Junwei and Jiang, Lu and Murphy, Kevin and Yu, Ting and Hauptmann, Alexander},
  booktitle={The IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)},
  month = {June},
  year={2020}
}
@inproceedings{liang2020simaug,
  title={SimAug: Learning Robust Representations from Simulation for Trajectory Prediction},
  author={Liang, Junwei and Jiang, Lu and Hauptmann, Alexander},
  booktitle={Proceedings of the European Conference on Computer Vision (ECCV)},
  month = {August},
  year={2020}
}

Introduction

This paper proposes the first multi-future pedestrian trajectory prediction dataset and a multi-future prediction method called Multiverse.

This paper studies the problem of predicting the distribution over multiple possible future paths of people as they move through various visual scenes. We make two main contributions. The first contribution is a new dataset called the Forking Paths Dataset, created in a realistic 3D simulator, which is based on real world trajectory data, and then extrapolated by human annotators to achieve different latent goals. This provides the first benchmark for quantitative evaluation of the models to predict multi-future trajectories.

The Forking Paths Dataset

  • Current dataset version: v1

  • Download links: Google Drive (sha256sum) / Baidu Pan (提取码: tpd7)

  • The dataset includes 3000 1920x1080 videos (750 human-annotated trajectory samples in 4 camera views) with bounding boxes and scene semantic segmentation ground truth. More notes and instructions about the dataset can be found here.

  • Instructions of how to add more human annotations, edit the scenes, recreate from real-world videos, or just simply to play with 3D simulator, can be found here.

  • Instructions of how to semi-automatically re-create real-world videos' scenarios with homography matrices into 3D simulation can be found here.

The Multiverse Model


Our second contribution is a new model to generate multiple plausible future trajectories, which contains novel designs of using multi-scale location encodings and convolutional RNNs over graphs. We refer to our model as Multiverse.

Dependencies

  • Python 2/3; TensorFlow-GPU >= 1.15.0

Pretrained Models

You can download pretrained models by running the script bash scripts/download_single_models.sh.

Testing and Visualization

Instructions for testing pretrained models can be found here.

Qualitative analysis between Social-GAN (left) and our model.

Training new models

Instructions for training new models can be found here.

Acknowledgments

The Forking Paths Dataset is created based on the CARLA Simulator and Unreal Engine 4.

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].