The-Learning-Machines / SBR

Licence: other
⌛ Introducing Self-Attention to Target Attentive Graph Neural Networks (AISP '22)

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to SBR

mtad-gat-pytorch
PyTorch implementation of MTAD-GAT (Multivariate Time-Series Anomaly Detection via Graph Attention Networks) by Zhao et. al (2020, https://arxiv.org/abs/2009.02040).
Stars: ✭ 85 (+286.36%)
Mutual labels:  attention, graph-neural-networks
how attentive are gats
Code for the paper "How Attentive are Graph Attention Networks?" (ICLR'2022)
Stars: ✭ 200 (+809.09%)
Mutual labels:  attention, graph-neural-networks
gnn-lspe
Source code for GNN-LSPE (Graph Neural Networks with Learnable Structural and Positional Representations), ICLR 2022
Stars: ✭ 165 (+650%)
Mutual labels:  attention, graph-neural-networks
InterGCN-ABSA
[COLING 2020] Jointly Learning Aspect-Focused and Inter-Aspect Relations with Graph Convolutional Networks for Aspect Sentiment Analysis
Stars: ✭ 41 (+86.36%)
Mutual labels:  graph-neural-networks
RL-based-Graph2Seq-for-NQG
Code & data accompanying the ICLR 2020 paper "Reinforcement Learning Based Graph-to-Sequence Model for Natural Question Generation"
Stars: ✭ 104 (+372.73%)
Mutual labels:  graph-neural-networks
SelfTask-GNN
Implementation of paper "Self-supervised Learning on Graphs:Deep Insights and New Directions"
Stars: ✭ 78 (+254.55%)
Mutual labels:  graph-neural-networks
well-classified-examples-are-underestimated
Code for the AAAI 2022 publication "Well-classified Examples are Underestimated in Classification with Deep Neural Networks"
Stars: ✭ 21 (-4.55%)
Mutual labels:  graph-neural-networks
AoA-pytorch
A Pytorch implementation of Attention on Attention module (both self and guided variants), for Visual Question Answering
Stars: ✭ 33 (+50%)
Mutual labels:  attention
ProteinGCN
ProteinGCN: Protein model quality assessment using Graph Convolutional Networks
Stars: ✭ 88 (+300%)
Mutual labels:  graph-neural-networks
Base-On-Relation-Method-Extract-News-DA-RNN-Model-For-Stock-Prediction--Pytorch
基於關聯式新聞提取方法之雙階段注意力機制模型用於股票預測
Stars: ✭ 33 (+50%)
Mutual labels:  attention
Walk-Transformer
From Random Walks to Transformer for Learning Node Embeddings (ECML-PKDD 2020) (In Pytorch and Tensorflow)
Stars: ✭ 26 (+18.18%)
Mutual labels:  graph-neural-networks
NTUA-slp-nlp
💻Speech and Natural Language Processing (SLP & NLP) Lab Assignments for ECE NTUA
Stars: ✭ 19 (-13.64%)
Mutual labels:  attention
gemnet pytorch
GemNet model in PyTorch, as proposed in "GemNet: Universal Directional Graph Neural Networks for Molecules" (NeurIPS 2021)
Stars: ✭ 80 (+263.64%)
Mutual labels:  graph-neural-networks
RNNSearch
An implementation of attention-based neural machine translation using Pytorch
Stars: ✭ 43 (+95.45%)
Mutual labels:  attention
keras cv attention models
Keras/Tensorflow attention models including beit,botnet,CMT,CoaT,CoAtNet,convnext,cotnet,davit,efficientdet,efficientnet,fbnet,gmlp,halonet,lcnet,levit,mlp-mixer,mobilevit,nfnets,regnet,resmlp,resnest,resnext,resnetd,swin,tinynet,uniformer,volo,wavemlp,yolor,yolox
Stars: ✭ 159 (+622.73%)
Mutual labels:  attention
deepsphere-cosmo-tf1
A spherical convolutional neural network for cosmology (TFv1).
Stars: ✭ 119 (+440.91%)
Mutual labels:  graph-neural-networks
Spectral-Designed-Graph-Convolutions
Codes for "Bridging the Gap Between Spectral and Spatial Domains in Graph Neural Networks" paper
Stars: ✭ 39 (+77.27%)
Mutual labels:  graph-neural-networks
H-GCN
[IJCAI 2019] Source code and datasets for "Hierarchical Graph Convolutional Networks for Semi-supervised Node Classification"
Stars: ✭ 103 (+368.18%)
Mutual labels:  graph-neural-networks
LibAUC
An End-to-End Machine Learning Library to Optimize AUC (AUROC, AUPRC).
Stars: ✭ 115 (+422.73%)
Mutual labels:  graph-neural-networks
interpretable-han-for-document-classification-with-keras
Keras implementation of hierarchical attention network for document classification with options to predict and present attention weights on both word and sentence level.
Stars: ✭ 18 (-18.18%)
Mutual labels:  attention

Introducing Self-Attention to Target Attentive Graph Neural Networks (arXiv)

[ Model Architecture ] [ Task ] [ Paper ] [ arXiv ]

PyTorch implementation of the model TAGNN++, presented in the paper "Introducing Self-Attention to Target Attentive Graph Neural Networks"

Accepted at AISP '22


Usage

python3 train.py [-h] [--dataset DATASET] [--defaults DEFAULTS] [--batchSize BATCHSIZE]
               [--hiddenSize HIDDENSIZE] [--nhead NHEAD] [--layer LAYER]
               [--feedforward FEEDFORWARD] [--epoch EPOCH] [--lr LR]
               [--lr_dc LR_DC] [--lr_dc_step LR_DC_STEP] [--l2 L2]
               [--patience PATIENCE] [--validation] 
               [--valid_portion VALID_PORTION]

Arguments:
  -h, --help            Description and Help Message
  --dataset DATASET     Name of the Dataset:
                        diginetica | yoochoose1_64
  --defaults DEFAULTS   Use default configuration   
  --batchSize BATCHSIZE
                        Batch size
  --hiddenSize HIDDENSIZE
                        Hidden state dimensions
  --epoch EPOCH         The number of epochs to train
  --lr LR               Set the Learning Rate
  --lr_dc LR_DC         Set the decay rate used with Learning rate
  --lr_dc_step LR_DC_STEP
                        Steps in Learning rate decay
  --l2 L2               Assign L2 Penalty
  --patience PATIENCE   Early stopping criterion
  --validation          validation
  --valid_portion VALID_PORTION
                        Portion of train-set to split into val-set

Dependencies

This code was developed with python3.6

Python (3.6.x)
PyTorch (1.7.x)
CUDA (10.2)
cuDNN (7.6.5)
networkx (2.5.1)
numpy (1.19.5)     

For original source of AGC, for further tweaks:

git clone https://github.com/vballoli/nfnets-pytorch.git   

Problem Statement Formulation

For further details, contact Sai Mitheran via Linkedin, or via email by clicking the icon below.


Reference

To cite our paper:

@article{mitheran2021mproved,
    title={Improved Representation Learning for Session-based Recommendation},
    author={Sai Mitheran and Abhinav Java and Surya Kant Sahu and Arshad Shaikh},
    year={2021},
    journal={arXiv preprint arXiv:2107.01516}
}

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].