All Projects → open-mmlab → Mmaction

open-mmlab / Mmaction

Licence: apache-2.0
An open-source toolbox for action understanding based on PyTorch

Programming Languages

python
139335 projects - #7 most used programming language
Cuda
1817 projects
C++
36643 projects - #6 most used programming language
shell
77523 projects
Dockerfile
14818 projects

Projects that are alternatives of or similar to Mmaction

Awesome Action Recognition
A curated list of action recognition and related area resources
Stars: ✭ 3,202 (+87.14%)
Mutual labels:  action-recognition, video-understanding, action-detection
TadTR
End-to-end Temporal Action Detection with Transformer. [Under review for a journal publication]
Stars: ✭ 55 (-96.79%)
Mutual labels:  action-recognition, temporal-action-detection, temporal-action-localization
MUSES
[CVPR 2021] Multi-shot Temporal Event Localization: a Benchmark
Stars: ✭ 51 (-97.02%)
Mutual labels:  action-recognition, temporal-action-detection, temporal-action-localization
Tdn
[CVPR 2021] TDN: Temporal Difference Networks for Efficient Action Recognition
Stars: ✭ 72 (-95.79%)
Mutual labels:  action-recognition, video-understanding
MTL-AQA
What and How Well You Performed? A Multitask Learning Approach to Action Quality Assessment [CVPR 2019]
Stars: ✭ 38 (-97.78%)
Mutual labels:  action-recognition, video-understanding
Materials-Temporal-Action-Detection
temporal action detection: benchmark results, features download etc.
Stars: ✭ 199 (-88.37%)
Mutual labels:  action-detection, temporal-action-detection
Awesome-Weakly-Supervised-Temporal-Action-Localization
A curated publication list on weakly-supervised temporal action localization
Stars: ✭ 65 (-96.2%)
Mutual labels:  temporal-action-detection, temporal-action-localization
DIN-Group-Activity-Recognition-Benchmark
A new codebase for Group Activity Recognition. It contains codes for ICCV 2021 paper: Spatio-Temporal Dynamic Inference Network for Group Activity Recognition and some other methods.
Stars: ✭ 26 (-98.48%)
Mutual labels:  action-recognition, video-understanding
DEAR
[ICCV 2021 Oral] Deep Evidential Action Recognition
Stars: ✭ 36 (-97.9%)
Mutual labels:  action-recognition, video-understanding
Mmaction2
OpenMMLab's Next Generation Video Understanding Toolbox and Benchmark
Stars: ✭ 684 (-60.02%)
Mutual labels:  action-recognition, video-understanding
Action Detection
temporal action detection with SSN
Stars: ✭ 597 (-65.11%)
Mutual labels:  action-recognition, video-understanding
Movienet Tools
Tools for movie and video research
Stars: ✭ 113 (-93.4%)
Mutual labels:  action-recognition, video-understanding
Paddlevideo
Comprehensive, latest, and deployable video deep learning algorithm, including video recognition, action localization, and temporal action detection tasks. It's a high-performance, light-weight codebase provides practical models for video understanding research and application
Stars: ✭ 218 (-87.26%)
Mutual labels:  action-recognition, video-understanding
Actionvlad
ActionVLAD for video action classification (CVPR 2017)
Stars: ✭ 217 (-87.32%)
Mutual labels:  action-recognition, video-understanding
Temporal Segment Networks
Code & Models for Temporal Segment Networks (TSN) in ECCV 2016
Stars: ✭ 1,287 (-24.78%)
Mutual labels:  action-recognition, video-understanding
Step
STEP: Spatio-Temporal Progressive Learning for Video Action Detection. CVPR'19 (Oral)
Stars: ✭ 196 (-88.54%)
Mutual labels:  action-recognition, video-understanding
Awesome Activity Prediction
Paper list of activity prediction and related area
Stars: ✭ 147 (-91.41%)
Mutual labels:  action-recognition, video-understanding
Video Understanding Dataset
A collection of recent video understanding datasets, under construction!
Stars: ✭ 387 (-77.38%)
Mutual labels:  action-recognition, video-understanding
Tsn Pytorch
Temporal Segment Networks (TSN) in PyTorch
Stars: ✭ 895 (-47.69%)
Mutual labels:  action-recognition, video-understanding
I3d finetune
TensorFlow code for finetuning I3D model on UCF101.
Stars: ✭ 128 (-92.52%)
Mutual labels:  action-recognition, video-understanding

MMAction

Introduction

MMAction is an open source toolbox for action understanding based on PyTorch. It is a part of the open-mmlab project developed by Multimedia Laboratory, CUHK.

Major Features

  • MMAction is capable of dealing with all of the tasks below.

    • action recognition from trimmed videos
    • temporal action detection (also known as action localization) in untrimmed videos
    • spatial-temporal action detection in untrimmed videos.
  • Support for various datasets

    Video datasets have emerging throughout the recent years and have greatly fostered the devlopment of this field. MMAction provides tools to deal with various datasets.

  • Support for multiple action understanding frameworks

    MMAction implements popular frameworks for action understanding:

    • For action recognition, various algorithms are implemented, including TSN, I3D, SlowFast, R(2+1)D, CSN.
    • For temporal action detection, we implement SSN.
    • For spatial temporal atomic action detection, a Fast-RCNN baseline is provided.
  • Modular design

    The tasks in human action understanding share some common aspects such as backbones, and long-term and short-term sampling schemes. Also, tasks can benefit from each other. For example, a better backbone for action recognition will bring performance gain for action detection. Modular design enables us to view action understanding in a more integrated perspective.

License

The project is release under the Apache 2.0 license.

Updates

OmniSource Model Release (22/08/2020)

  • We release several models of our work OmniSource. These models are jointly trained with Kinetics-400 and OmniSourced web dataset. Those models are of good performance (Top1 Accuracy: 75.7% for 3-segment TSN and 80.4% for SlowOnly on Kinetics-400 val) and the learned representation transfer well to other tasks.

v0.2.0 (15/03/2020)

  • We build a diversified modelzoo for action recognition, which include popular algorithms (TSN, I3D, SlowFast, R(2+1)D, CSN). The performance is aligned with or better than the original papers.

v0.1.0 (19/06/2019)

  • MMAction is online!

Model zoo

Results and reference models are available in the model zoo.

Installation

Please refer to INSTALL.md for installation.

Update: for Docker installation, Please refer to DOCKER.md for using docker for this project.

Data preparation

Please refer to DATASET.md for a general knowledge of data preparation. Detailed documents for the supported datasets are available in data_tools/.

Get started

Please refer to GETTING_STARTED.md for detailed examples and abstract usage.

Contributing

We appreciate all contributions to improve MMAction. Please refer to CONTRUBUTING.md for the contributing guideline.

Citation

If you use our codebase or models in your research, please cite this work. We will release a technical report later.

@misc{mmaction2019,
  author =       {Yue Zhao, Yuanjun Xiong, Dahua Lin},
  title =        {MMAction},
  howpublished = {\url{https://github.com/open-mmlab/mmaction}},
  year =         {2019}
}

Contact

If you have any question, please file an issue or contact the author:

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].