brainGraphGraph theory analysis of brain MRI data
Stars: ✭ 136 (+300%)
spatio-temporal-brainA Deep Graph Neural Network Architecture for Modelling Spatio-temporal Dynamics in rs-fMRI Data
Stars: ✭ 22 (-35.29%)
visualqcVisualQC : assistive tool to ease the quality control workflow of neuroimaging data.
Stars: ✭ 56 (+64.71%)
ReFineOfficial code of "Towards Multi-Grained Explainability for Graph Neural Networks" (2021 NeurIPS)
Stars: ✭ 40 (+17.65%)
fmriflowsfmriflows is a consortium of many (dependent) fMRI analysis pipelines, including anatomical and functional pre-processing, univariate 1st and 2nd-level analysis, as well as multivariate pattern analysis.
Stars: ✭ 40 (+17.65%)
mtad-gat-pytorchPyTorch implementation of MTAD-GAT (Multivariate Time-Series Anomaly Detection via Graph Attention Networks) by Zhao et. al (2020, https://arxiv.org/abs/2009.02040).
Stars: ✭ 85 (+150%)
GNNs-in-Network-NeuroscienceA review of papers proposing novel GNN methods with application to brain connectivity published in 2017-2020.
Stars: ✭ 92 (+170.59%)
Revisiting-Contrastive-SSLRevisiting Contrastive Methods for Unsupervised Learning of Visual Representations. [NeurIPS 2021]
Stars: ✭ 81 (+138.24%)
gnn-lspeSource code for GNN-LSPE (Graph Neural Networks with Learnable Structural and Positional Representations), ICLR 2022
Stars: ✭ 165 (+385.29%)
PDNThe official PyTorch implementation of "Pathfinder Discovery Networks for Neural Message Passing" (WebConf '21)
Stars: ✭ 44 (+29.41%)
connectomemapper3Connectome Mapper 3 is a BIDS App that implements full anatomical, diffusion, resting/state functional MRI, and recently EEG processing pipelines, from raw T1 / DWI / BOLD , and preprocessed EEG data to multi-resolution brain parcellation with corresponding connection matrices.
Stars: ✭ 45 (+32.35%)
clinicaSoftware platform for clinical neuroimaging studies
Stars: ✭ 153 (+350%)
BasicGNNTrackingThis shows a basic implementation of the global nearest neighbour (GNN) multi target Tracker. Kalman filter is used for Tracking and Auction Algorithm for determining the assignment of measurments to filters.
Stars: ✭ 36 (+5.88%)
BGCNA Tensorflow implementation of "Bayesian Graph Convolutional Neural Networks" (AAAI 2019).
Stars: ✭ 129 (+279.41%)
mllpThe code of AAAI 2020 paper "Transparent Classification with Multilayer Logical Perceptrons and Random Binarization".
Stars: ✭ 15 (-55.88%)
brainlitMethod container for computational neuroscience on brains.
Stars: ✭ 20 (-41.18%)
clinicadlFramework for the reproducible processing of neuroimaging data with deep learning methods
Stars: ✭ 114 (+235.29%)
BrainMaGeBrain extraction in presence of abnormalities, using single and multiple MRI modalities
Stars: ✭ 23 (-32.35%)
AttnSleep[IEEE TNSRE] "An Attention-based Deep Learning Approach for Sleep Stage Classification with Single-Channel EEG"
Stars: ✭ 76 (+123.53%)
RecycleNetAttentional Learning of Trash Classification
Stars: ✭ 23 (-32.35%)
Hyper-SAGNNhypergraph representation learning, graph neural network
Stars: ✭ 53 (+55.88%)
chatbot一个基于深度学习的中文聊天机器人,这里有详细的教程与代码,每份代码都有详细的注释,作为学习是美好的选择。A Chinese chatbot based on deep learning.
Stars: ✭ 94 (+176.47%)
responsible-ai-toolboxThis project provides responsible AI user interfaces for Fairlearn, interpret-community, and Error Analysis, as well as foundational building blocks that they rely on.
Stars: ✭ 615 (+1708.82%)
GNN-Recommender-SystemsAn index of recommendation algorithms that are based on Graph Neural Networks.
Stars: ✭ 505 (+1385.29%)
EBIM-NLIEnhanced BiLSTM Inference Model for Natural Language Inference
Stars: ✭ 24 (-29.41%)
gnn-re-rankingA real-time GNN-based method. Understanding Image Retrieval Re-Ranking: A Graph Neural Network Perspective
Stars: ✭ 64 (+88.24%)
h-transformer-1dImplementation of H-Transformer-1D, Hierarchical Attention for Sequence Learning
Stars: ✭ 121 (+255.88%)
transformerA PyTorch Implementation of "Attention Is All You Need"
Stars: ✭ 28 (-17.65%)
egnn-pytorchImplementation of E(n)-Equivariant Graph Neural Networks, in Pytorch
Stars: ✭ 249 (+632.35%)
External-Attention-pytorch🍀 Pytorch implementation of various Attention Mechanisms, MLP, Re-parameter, Convolution, which is helpful to further understand papers.⭐⭐⭐
Stars: ✭ 7,344 (+21500%)
contextual-aiContextual AI adds explainability to different stages of machine learning pipelines - data, training, and inference - thereby addressing the trust gap between such ML systems and their users. It does not refer to a specific algorithm or ML method — instead, it takes a human-centric view and approach to AI.
Stars: ✭ 81 (+138.24%)
TRAR-VQA[ICCV 2021] TRAR: Routing the Attention Spans in Transformers for Visual Question Answering -- Official Implementation
Stars: ✭ 49 (+44.12%)
deep-explanation-penalizationCode for using CDEP from the paper "Interpretations are useful: penalizing explanations to align neural networks with prior knowledge" https://arxiv.org/abs/1909.13584
Stars: ✭ 110 (+223.53%)
socialRLCode and data for Zhang, Lengersdorff et al. (2020)
Stars: ✭ 19 (-44.12%)
bidskitUtility functions for working with DICOM and BIDS neuroimaging data
Stars: ✭ 52 (+52.94%)
hierarchical-dnn-interpretationsUsing / reproducing ACD from the paper "Hierarchical interpretations for neural network predictions" 🧠 (ICLR 2019)
Stars: ✭ 110 (+223.53%)
nidbNeuroInformatics Database
Stars: ✭ 26 (-23.53%)
GCLList of Publications in Graph Contrastive Learning
Stars: ✭ 25 (-26.47%)
chunkflowCompose chunk operators to create a pipeline for local or distributed petabyte-scale computation
Stars: ✭ 36 (+5.88%)
ENIGMAThe ENIGMA Toolbox is an open-source repository for accessing 100+ ENIGMA statistical maps, visualizing cortical and subcortical surface data, and relating neuroimaging findings to micro- and macroscale brain organization. 🤠
Stars: ✭ 66 (+94.12%)
giftiMATLAB/Octave GIfTI Library
Stars: ✭ 16 (-52.94%)
MixGCFMixGCF: An Improved Training Method for Graph Neural Network-based Recommender Systems, KDD2021
Stars: ✭ 73 (+114.71%)
dgcnnClean & Documented TF2 implementation of "An end-to-end deep learning architecture for graph classification" (M. Zhang et al., 2018).
Stars: ✭ 21 (-38.24%)
mne-bidsMNE-BIDS is a Python package that allows you to read and write BIDS-compatible datasets with the help of MNE-Python.
Stars: ✭ 88 (+158.82%)
CrabNetPredict materials properties using only the composition information!
Stars: ✭ 57 (+67.65%)
Transformer-MM-Explainability[ICCV 2021- Oral] Official PyTorch implementation for Generic Attention-model Explainability for Interpreting Bi-Modal and Encoder-Decoder Transformers, a novel method to visualize any Transformer-based network. Including examples for DETR, VQA.
Stars: ✭ 484 (+1323.53%)
mmgnn textvqaA Pytorch implementation of CVPR 2020 paper: Multi-Modal Graph Neural Network for Joint Reasoning on Vision and Scene Text
Stars: ✭ 41 (+20.59%)
SuperGAT[ICLR 2021] How to Find Your Friendly Neighborhood: Graph Attention Design with Self-Supervision
Stars: ✭ 122 (+258.82%)
DeepMoveCodes for WWW'18 Paper-DeepMove: Predicting Human Mobility with Attentional Recurrent Network
Stars: ✭ 120 (+252.94%)
DeepLearningReadingDeep Learning and Machine Learning mini-projects. Current Project: Deepmind Attentive Reader (rc-data)
Stars: ✭ 78 (+129.41%)