All Projects → jedyang97 → MTAG

jedyang97 / MTAG

Licence: MIT License
Code for NAACL 2021 paper: MTAG: Modal-Temporal Attention Graph for Unaligned Human Multimodal Language Sequences

Programming Languages

Jupyter Notebook
11667 projects
python
139335 projects - #7 most used programming language
shell
77523 projects

Projects that are alternatives of or similar to MTAG

awesome-graph-explainability-papers
Papers about explainability of GNNs
Stars: ✭ 153 (+565.22%)
Mutual labels:  graph-neural-networks
SelfTask-GNN
Implementation of paper "Self-supervised Learning on Graphs:Deep Insights and New Directions"
Stars: ✭ 78 (+239.13%)
Mutual labels:  graph-neural-networks
PathCon
Combining relational context and relational paths for knowledge graph completion
Stars: ✭ 94 (+308.7%)
Mutual labels:  graph-neural-networks
InterGCN-ABSA
[COLING 2020] Jointly Learning Aspect-Focused and Inter-Aspect Relations with Graph Convolutional Networks for Aspect Sentiment Analysis
Stars: ✭ 41 (+78.26%)
Mutual labels:  graph-neural-networks
H-GCN
[IJCAI 2019] Source code and datasets for "Hierarchical Graph Convolutional Networks for Semi-supervised Node Classification"
Stars: ✭ 103 (+347.83%)
Mutual labels:  graph-neural-networks
Spectral-Designed-Graph-Convolutions
Codes for "Bridging the Gap Between Spectral and Spatial Domains in Graph Neural Networks" paper
Stars: ✭ 39 (+69.57%)
Mutual labels:  graph-neural-networks
sdn-nfv-papers
This is a paper list about Resource Allocation in Network Functions Virtualization (NFV) and Software-Defined Networking (SDN).
Stars: ✭ 40 (+73.91%)
Mutual labels:  graph-neural-networks
DiGCN
Implement of DiGCN, NeurIPS-2020
Stars: ✭ 25 (+8.7%)
Mutual labels:  graph-neural-networks
Walk-Transformer
From Random Walks to Transformer for Learning Node Embeddings (ECML-PKDD 2020) (In Pytorch and Tensorflow)
Stars: ✭ 26 (+13.04%)
Mutual labels:  graph-neural-networks
SBR
⌛ Introducing Self-Attention to Target Attentive Graph Neural Networks (AISP '22)
Stars: ✭ 22 (-4.35%)
Mutual labels:  graph-neural-networks
RL-based-Graph2Seq-for-NQG
Code & data accompanying the ICLR 2020 paper "Reinforcement Learning Based Graph-to-Sequence Model for Natural Question Generation"
Stars: ✭ 104 (+352.17%)
Mutual labels:  graph-neural-networks
LibAUC
An End-to-End Machine Learning Library to Optimize AUC (AUROC, AUPRC).
Stars: ✭ 115 (+400%)
Mutual labels:  graph-neural-networks
ProteinGCN
ProteinGCN: Protein model quality assessment using Graph Convolutional Networks
Stars: ✭ 88 (+282.61%)
Mutual labels:  graph-neural-networks
deepsphere-cosmo-tf1
A spherical convolutional neural network for cosmology (TFv1).
Stars: ✭ 119 (+417.39%)
Mutual labels:  graph-neural-networks
AGCN
No description or website provided.
Stars: ✭ 17 (-26.09%)
Mutual labels:  graph-neural-networks
OpenHGNN
This is an open-source toolkit for Heterogeneous Graph Neural Network(OpenHGNN) based on DGL.
Stars: ✭ 264 (+1047.83%)
Mutual labels:  graph-neural-networks
gemnet pytorch
GemNet model in PyTorch, as proposed in "GemNet: Universal Directional Graph Neural Networks for Molecules" (NeurIPS 2021)
Stars: ✭ 80 (+247.83%)
Mutual labels:  graph-neural-networks
DIN-Group-Activity-Recognition-Benchmark
A new codebase for Group Activity Recognition. It contains codes for ICCV 2021 paper: Spatio-Temporal Dynamic Inference Network for Group Activity Recognition and some other methods.
Stars: ✭ 26 (+13.04%)
Mutual labels:  graph-neural-networks
mtad-gat-pytorch
PyTorch implementation of MTAD-GAT (Multivariate Time-Series Anomaly Detection via Graph Attention Networks) by Zhao et. al (2020, https://arxiv.org/abs/2009.02040).
Stars: ✭ 85 (+269.57%)
Mutual labels:  graph-neural-networks
well-classified-examples-are-underestimated
Code for the AAAI 2022 publication "Well-classified Examples are Underestimated in Classification with Deep Neural Networks"
Stars: ✭ 21 (-8.7%)
Mutual labels:  graph-neural-networks

Paper: https://aclanthology.org/2021.naacl-main.79

Talk: https://www.youtube.com/watch?v=b0UogJP4U5I

MTAG (Modal-Temporal Attention Graph) is a GNN-based machine learning framework that can learn fusion and alignment for unaligned multimodal sequences.

Our code is written as an extension to the awesome PyTorch Geometric library. Users are encouraged to read their installation guide and documentations to understand the basics.

Our main contributions include:

  • A graph builder to construct graphs with modal and temporal edges.
  • A new GNN convolution operation called MTGATConv that uses distinct attentions for edges with distinct modality and temporal ordering. It also transforms each node based on its modality type. It is like a combination of RGCNConv and GATConv with an efficient implementation. We hope this operation can be inlcuded into PyTorch Geometric as a standard operation.
  • A TopK pooling operation to prune edges with low attention weights.

Installation

Please refer to the requirement.txt for setup.

Dataset Preperation

Download the following datasets (please copy and paste the URL to browser, as clicking the link might not work):

and put them into a desired folder (.e.g. <dataroot>). Then specify in run.sh the folder containing the data of the desired dataset. For example:

python main.py \
...
--dataroot <dataroot>
...

Running Example

bash run.sh

To visualize the edges:

jupyter notebook network_inference_visualize.ipynb

Hyperparameters

A more comprehensive hyperparameter list (along with each setting's performance we obtained) can be found in this Google Sheet. For any parameters that are not specified here, we used the default values in main.py.

Citation

@inproceedings{yang-etal-2021-mtag,
    title = "{MTAG}: Modal-Temporal Attention Graph for Unaligned Human Multimodal Language Sequences",
    author = "Yang, Jianing  and
      Wang, Yongxin  and
      Yi, Ruitao  and
      Zhu, Yuying  and
      Rehman, Azaan  and
      Zadeh, Amir  and
      Poria, Soujanya  and
      Morency, Louis-Philippe",
    booktitle = "Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies",
    month = jun,
    year = "2021",
    address = "Online",
    publisher = "Association for Computational Linguistics",
    url = "https://www.aclweb.org/anthology/2021.naacl-main.79",
    pages = "1009--1021",
    abstract = "Human communication is multimodal in nature; it is through multiple modalities such as language, voice, and facial expressions, that opinions and emotions are expressed. Data in this domain exhibits complex multi-relational and temporal interactions. Learning from this data is a fundamentally challenging research problem. In this paper, we propose Modal-Temporal Attention Graph (MTAG). MTAG is an interpretable graph-based neural model that provides a suitable framework for analyzing multimodal sequential data. We first introduce a procedure to convert unaligned multimodal sequence data into a graph with heterogeneous nodes and edges that captures the rich interactions across modalities and through time. Then, a novel graph fusion operation, called MTAG fusion, along with a dynamic pruning and read-out technique, is designed to efficiently process this modal-temporal graph and capture various interactions. By learning to focus only on the important interactions within the graph, MTAG achieves state-of-the-art performance on multimodal sentiment analysis and emotion recognition benchmarks, while utilizing significantly fewer model parameters.",
}
Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].