All Projects → anita-hu → MSAF

anita-hu / MSAF

Licence: MIT license
Offical implementation of paper "MSAF: Multimodal Split Attention Fusion"

Programming Languages

python
139335 projects - #7 most used programming language
shell
77523 projects

Projects that are alternatives of or similar to MSAF

Self-Supervised-Embedding-Fusion-Transformer
The code for our IEEE ACCESS (2020) paper Multimodal Emotion Recognition with Transformer-Based Self Supervised Feature Fusion.
Stars: ✭ 57 (+21.28%)
Mutual labels:  multimodal-sentiment-analysis, multimodal-deep-learning, multimodal-emotion-recognition
Action-Localization
Action-Localization, Atomic Visual Actions (AVA) Dataset
Stars: ✭ 22 (-53.19%)
Mutual labels:  action-recognition, ntu-rgbd
vista-net
Code for the paper "VistaNet: Visual Aspect Attention Network for Multimodal Sentiment Analysis", AAAI'19
Stars: ✭ 67 (+42.55%)
Mutual labels:  multimodal-learning, multimodal-sentiment-analysis
BBFN
This repository contains the implementation of the paper -- Bi-Bimodal Modality Fusion for Correlation-Controlled Multimodal Sentiment Analysis
Stars: ✭ 42 (-10.64%)
Mutual labels:  multimodal-sentiment-analysis, multimodal-deep-learning
slp
Utils and modules for Speech Language and Multimodal processing using pytorch and pytorch lightning
Stars: ✭ 17 (-63.83%)
Mutual labels:  multimodal-learning, multimodal-deep-learning
pose2action
experiments on classifying actions using poses
Stars: ✭ 24 (-48.94%)
Mutual labels:  action-recognition, ntu-rgbd
Robust-Deep-Learning-Pipeline
Deep Convolutional Bidirectional LSTM for Complex Activity Recognition with Missing Data. Human Activity Recognition Challenge. Springer SIST (2020)
Stars: ✭ 20 (-57.45%)
Mutual labels:  action-recognition, multimodal-deep-learning
Alphaction
Spatio-Temporal Action Localization System
Stars: ✭ 221 (+370.21%)
Mutual labels:  action-recognition
temporal-binding-network
Implementation of "EPIC-Fusion: Audio-Visual Temporal Binding for Egocentric Action Recognition, ICCV, 2019" in PyTorch
Stars: ✭ 95 (+102.13%)
Mutual labels:  action-recognition
Action recognition zoo
Codes for popular action recognition models, verified on the something-something data set.
Stars: ✭ 227 (+382.98%)
Mutual labels:  action-recognition
MiCT-Net-PyTorch
Video Recognition using Mixed Convolutional Tube (MiCT) on PyTorch with a ResNet backbone
Stars: ✭ 48 (+2.13%)
Mutual labels:  action-recognition
temporal-ssl
Video Representation Learning by Recognizing Temporal Transformations. In ECCV, 2020.
Stars: ✭ 46 (-2.13%)
Mutual labels:  action-recognition
Paddlevideo
Comprehensive, latest, and deployable video deep learning algorithm, including video recognition, action localization, and temporal action detection tasks. It's a high-performance, light-weight codebase provides practical models for video understanding research and application
Stars: ✭ 218 (+363.83%)
Mutual labels:  action-recognition
Lintel
A Python module to decode video frames directly, using the FFmpeg C API.
Stars: ✭ 240 (+410.64%)
Mutual labels:  action-recognition
Keras-for-Co-occurrence-Feature-Learning-from-Skeleton-Data-for-Action-Recognition
Keras implementation for Co-occurrence-Feature-Learning-from-Skeleton-Data-for-Action-Recognition
Stars: ✭ 44 (-6.38%)
Mutual labels:  action-recognition
Ms G3d
[CVPR 2020 Oral] PyTorch implementation of "Disentangling and Unifying Graph Convolutions for Skeleton-Based Action Recognition"
Stars: ✭ 225 (+378.72%)
Mutual labels:  action-recognition
C3D-tensorflow
Action recognition with C3D network implemented in tensorflow
Stars: ✭ 34 (-27.66%)
Mutual labels:  action-recognition
Ican
[BMVC 2018] iCAN: Instance-Centric Attention Network for Human-Object Interaction Detection
Stars: ✭ 225 (+378.72%)
Mutual labels:  action-recognition
pykale
Knowledge-Aware machine LEarning (KALE): accessible machine learning from multiple sources for interdisciplinary research, part of the 🔥PyTorch ecosystem
Stars: ✭ 381 (+710.64%)
Mutual labels:  multimodal-learning
weakly-action-localization
No description or website provided.
Stars: ✭ 30 (-36.17%)
Mutual labels:  action-recognition

MSAF: Multimodal Split Attention Fusion

arXiv PWC

Code for the paper MSAF: Multimodal Split Attention Fusion. This is our implementation of the MSAF module and the three MSAF-powered multimodal networks.

If you use this code, please cite our paper:

@misc{su2020msaf,
      title={MSAF: Multimodal Split Attention Fusion}, 
      author={Lang Su and Chuqing Hu and Guofa Li and Dongpu Cao},
      year={2020},
      eprint={2012.07175},
      archivePrefix={arXiv},
      primaryClass={cs.CV}
}

Installation

Clone this repo along with submodules

git clone --recurse-submodules https://github.com/anita-hu/MSAF.git

Install dependencies

Method 1: Using environment.yml (installs dependencies for all three datasets)

With this method, you can skip dependency installation steps from the dataset specific README files

conda env create -f environment.yml

Method 2: Without environment.yml

This code was developed with Python 3.6, PyTorch 1.7.1 in Ubuntu 20.04.

  • Basic dependencies (needed for all datasets): Pytorch, Tensorboard
  • Dataset specific dependencies: see README file in each dataset folder

Usage

  • The MSAF module is implemented in MSAF.py
  • The README file in each dataset folder has details on data preprocessing, training and evaluation (pretrained weights are available)
Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].