resolutions-2019A list of data mining and machine learning papers that I implemented in 2019.
Stars: ✭ 19 (-99.56%)
FEATHERThe reference implementation of FEATHER from the CIKM '20 paper "Characteristic Functions on Graphs: Birds of a Feather, from Statistical Descriptors to Parametric Models".
Stars: ✭ 34 (-99.21%)
FSCNMFAn implementation of "Fusing Structure and Content via Non-negative Matrix Factorization for Embedding Information Networks".
Stars: ✭ 16 (-99.63%)
EulerA distributed graph deep learning framework.
Stars: ✭ 2,701 (-37.32%)
walkletsA lightweight implementation of Walklets from "Don't Walk Skip! Online Learning of Multi-scale Network Embeddings" (ASONAM 2017).
Stars: ✭ 94 (-97.82%)
RolXAn alternative implementation of Recursive Feature and Role Extraction (KDD11 & KDD12)
Stars: ✭ 52 (-98.79%)
PDNThe official PyTorch implementation of "Pathfinder Discovery Networks for Neural Message Passing" (WebConf '21)
Stars: ✭ 44 (-98.98%)
dgcnnClean & Documented TF2 implementation of "An end-to-end deep learning architecture for graph classification" (M. Zhang et al., 2018).
Stars: ✭ 21 (-99.51%)
GE-FSGGraph Embedding via Frequent Subgraphs
Stars: ✭ 39 (-99.09%)
REGALRepresentation learning-based graph alignment based on implicit matrix factorization and structural embeddings
Stars: ✭ 78 (-98.19%)
NMFADMMA sparsity aware implementation of "Alternating Direction Method of Multipliers for Non-Negative Matrix Factorization with the Beta-Divergence" (ICASSP 2014).
Stars: ✭ 39 (-99.09%)
SelfGNNA PyTorch implementation of "SelfGNN: Self-supervised Graph Neural Networks without explicit negative sampling" paper, which appeared in The International Workshop on Self-Supervised Learning for the Web (SSL'21) @ the Web Conference 2021 (WWW'21).
Stars: ✭ 24 (-99.44%)
graphkit-learnA python package for graph kernels, graph edit distances, and graph pre-image problem.
Stars: ✭ 87 (-97.98%)
M-NMFAn implementation of "Community Preserving Network Embedding" (AAAI 2017)
Stars: ✭ 119 (-97.24%)
GatGraph Attention Networks (https://arxiv.org/abs/1710.10903)
Stars: ✭ 2,229 (-48.27%)
CoVA-Web-Object-DetectionA Context-aware Visual Attention-based training pipeline for Object Detection from a Webpage screenshot!
Stars: ✭ 18 (-99.58%)
GraphembeddingImplementation and experiments of graph embedding algorithms.
Stars: ✭ 2,461 (-42.89%)
PygatPytorch implementation of the Graph Attention Network model by Veličković et. al (2017, https://arxiv.org/abs/1710.10903)
Stars: ✭ 1,853 (-57%)
OpenANEOpenANE: the first Open source framework specialized in Attributed Network Embedding. The related paper was accepted by Neurocomputing. https://doi.org/10.1016/j.neucom.2020.05.080
Stars: ✭ 39 (-99.09%)
awesome-efficient-gnnCode and resources on scalable and efficient Graph Neural Networks
Stars: ✭ 498 (-88.44%)
GNN-Recommender-SystemsAn index of recommendation algorithms that are based on Graph Neural Networks.
Stars: ✭ 505 (-88.28%)
TriDNRTri-Party Deep Network Representation, IJCAI-16
Stars: ✭ 72 (-98.33%)
QGNNQuaternion Graph Neural Networks (ACML 2021) (Pytorch and Tensorflow)
Stars: ✭ 31 (-99.28%)
linformerImplementation of Linformer for Pytorch
Stars: ✭ 119 (-97.24%)
transganformerImplementation of TransGanFormer, an all-attention GAN that combines the finding from the recent GanFormer and TransGan paper
Stars: ✭ 137 (-96.82%)
mtad-gat-pytorchPyTorch implementation of MTAD-GAT (Multivariate Time-Series Anomaly Detection via Graph Attention Networks) by Zhao et. al (2020, https://arxiv.org/abs/2009.02040).
Stars: ✭ 85 (-98.03%)
TransformerA TensorFlow Implementation of the Transformer: Attention Is All You Need
Stars: ✭ 3,646 (-15.39%)
Vit PytorchImplementation of Vision Transformer, a simple way to achieve SOTA in vision classification with only a single transformer encoder, in Pytorch
Stars: ✭ 7,199 (+67.07%)
ADL2019Applied Deep Learning (2019 Spring) @ NTU
Stars: ✭ 20 (-99.54%)
vista-netCode for the paper "VistaNet: Visual Aspect Attention Network for Multimodal Sentiment Analysis", AAAI'19
Stars: ✭ 67 (-98.45%)
Multi Scale AttentionCode for our paper "Multi-scale Guided Attention for Medical Image Segmentation"
Stars: ✭ 281 (-93.48%)
Neural spEnd-to-end ASR/LM implementation with PyTorch
Stars: ✭ 408 (-90.53%)
Keras GatKeras implementation of the graph attention networks (GAT) by Veličković et al. (2017; https://arxiv.org/abs/1710.10903)
Stars: ✭ 334 (-92.25%)
Timesformer PytorchImplementation of TimeSformer from Facebook AI, a pure attention-based solution for video classification
Stars: ✭ 225 (-94.78%)
pynmta simple and complete pytorch implementation of neural machine translation system
Stars: ✭ 13 (-99.7%)
Attention一些不同的Attention机制代码
Stars: ✭ 17 (-99.61%)
TransformerA Pytorch Implementation of "Attention is All You Need" and "Weighted Transformer Network for Machine Translation"
Stars: ✭ 271 (-93.71%)
co-attentionPytorch implementation of "Dynamic Coattention Networks For Question Answering"
Stars: ✭ 54 (-98.75%)
Yolo Multi Backbones AttentionModel Compression—YOLOv3 with multi lightweight backbones(ShuffleNetV2 HuaWei GhostNet), attention, prune and quantization
Stars: ✭ 317 (-92.64%)
CIKM18-LCVACode for CIKM'18 paper, Linked Causal Variational Autoencoder for Inferring Paired Spillover Effects.
Stars: ✭ 13 (-99.7%)
GraphTSNEPyTorch Implementation of GraphTSNE, ICLR’19
Stars: ✭ 113 (-97.38%)
Da Rnn📃 **Unofficial** PyTorch Implementation of DA-RNN (arXiv:1704.02971)
Stars: ✭ 256 (-94.06%)
mvGAEDrug Similarity Integration Through Attentive Multi-view Graph Auto-Encoders (IJCAI 2018)
Stars: ✭ 27 (-99.37%)
Pytorch Original TransformerMy implementation of the original transformer model (Vaswani et al.). I've additionally included the playground.py file for visualizing otherwise seemingly hard concepts. Currently included IWSLT pretrained models.
Stars: ✭ 411 (-90.46%)
PaperrobotCode for PaperRobot: Incremental Draft Generation of Scientific Ideas
Stars: ✭ 372 (-91.37%)
Seq2seq chatbot基于seq2seq模型的简单对话系统的tf实现,具有embedding、attention、beam_search等功能,数据集是Cornell Movie Dialogs
Stars: ✭ 308 (-92.85%)
MirnetOfficial repository for "Learning Enriched Features for Real Image Restoration and Enhancement" (ECCV 2020). SOTA results for image denoising, super-resolution, and image enhancement.
Stars: ✭ 247 (-94.27%)
kafboxA Matlab benchmarking toolbox for kernel adaptive filtering
Stars: ✭ 70 (-98.38%)
GOSHAn ultra-fast, GPU-based large graph embedding algorithm utilizing a novel coarsening algorithm requiring not more than a single GPU.
Stars: ✭ 12 (-99.72%)
MachineLearningSeriesVídeos e códigos do Universo Discreto ensinando o fundamental de Machine Learning em Python. Para mais detalhes, acompanhar a playlist listada.
Stars: ✭ 20 (-99.54%)
GRACE[GRL+ @ ICML 2020] PyTorch implementation for "Deep Graph Contrastive Representation Learning" (https://arxiv.org/abs/2006.04131v2)
Stars: ✭ 144 (-96.66%)
MoChA-pytorchPyTorch Implementation of "Monotonic Chunkwise Attention" (ICLR 2018)
Stars: ✭ 65 (-98.49%)