All Projects → Self Attention Cv → Similar Projects or Alternatives

1536 Open source projects that are alternatives of or similar to Self Attention Cv

Isab Pytorch
An implementation of (Induced) Set Attention Block, from the Set Transformers paper
Stars: ✭ 21 (-89.95%)
CrabNet
Predict materials properties using only the composition information!
Stars: ✭ 57 (-72.73%)
Pytorch Original Transformer
My implementation of the original transformer model (Vaswani et al.). I've additionally included the playground.py file for visualizing otherwise seemingly hard concepts. Currently included IWSLT pretrained models.
Stars: ✭ 411 (+96.65%)
Se3 Transformer Pytorch
Implementation of SE3-Transformers for Equivariant Self-Attention, in Pytorch. This specific repository is geared towards integration with eventual Alphafold2 replication.
Stars: ✭ 73 (-65.07%)
Lambda Networks
Implementation of LambdaNetworks, a new approach to image recognition that reaches SOTA with less compute
Stars: ✭ 1,497 (+616.27%)
h-transformer-1d
Implementation of H-Transformer-1D, Hierarchical Attention for Sequence Learning
Stars: ✭ 121 (-42.11%)
Routing Transformer
Fully featured implementation of Routing Transformer
Stars: ✭ 149 (-28.71%)
visualization
a collection of visualization function
Stars: ✭ 189 (-9.57%)
Linear Attention Transformer
Transformer based on a variant of attention that is linear complexity in respect to sequence length
Stars: ✭ 205 (-1.91%)
Global Self Attention Network
A Pytorch implementation of Global Self-Attention Network, a fully-attention backbone for vision tasks
Stars: ✭ 64 (-69.38%)
Neural sp
End-to-end ASR/LM implementation with PyTorch
Stars: ✭ 408 (+95.22%)
Performer Pytorch
An implementation of Performer, a linear attention-based transformer, in Pytorch
Stars: ✭ 546 (+161.24%)
Cell Detr
Official and maintained implementation of the paper Attention-Based Transformers for Instance Segmentation of Cells in Microstructures [BIBM 2020].
Stars: ✭ 26 (-87.56%)
Mutual labels:  attention, transformer
Letslearnai.github.io
Lets Learn AI
Stars: ✭ 33 (-84.21%)
Machine Learning From Scratch
Succinct Machine Learning algorithm implementations from scratch in Python, solving real-world problems (Notebooks and Book). Examples of Logistic Regression, Linear Regression, Decision Trees, K-means clustering, Sentiment Analysis, Recommender Systems, Neural Networks and Reinforcement Learning.
Stars: ✭ 42 (-79.9%)
Deeplearning Nlp Models
A small, interpretable codebase containing the re-implementation of a few "deep" NLP models in PyTorch. Colab notebooks to run with GPUs. Models: word2vec, CNNs, transformer, gpt.
Stars: ✭ 64 (-69.38%)
Mutual labels:  attention, transformer
Php Ml
PHP-ML - Machine Learning library for PHP
Stars: ✭ 7,900 (+3679.9%)
Mixture Of Experts
A Pytorch implementation of Sparsely-Gated Mixture of Experts, for massively increasing the parameter count of language models
Stars: ✭ 68 (-67.46%)
Neat Vision
Neat (Neural Attention) Vision, is a visualization tool for the attention mechanisms of deep-learning models for Natural Language Processing (NLP) tasks. (framework-agnostic)
Stars: ✭ 213 (+1.91%)
Mutual labels:  attention-mechanism, attention
Point Transformer Pytorch
Implementation of the Point Transformer layer, in Pytorch
Stars: ✭ 199 (-4.78%)
Attend infer repeat
A Tensorfflow implementation of Attend, Infer, Repeat
Stars: ✭ 82 (-60.77%)
Mutual labels:  attention-mechanism, attention
Papers Literature Ml Dl Rl Ai
Highly cited and useful papers related to machine learning, deep learning, AI, game theory, reinforcement learning
Stars: ✭ 1,341 (+541.63%)
Multiturndialogzoo
Multi-turn dialogue baselines written in PyTorch
Stars: ✭ 106 (-49.28%)
Mutual labels:  attention, transformer
Hackerrank
This is the Repository where you can find all the solution of the Problems which you solve on competitive platforms mainly HackerRank and HackerEarth
Stars: ✭ 68 (-67.46%)
Njunmt Tf
An open-source neural machine translation system developed by Natural Language Processing Group, Nanjing University.
Stars: ✭ 97 (-53.59%)
Mutual labels:  attention, transformer
Overlappredator
[CVPR 2021, Oral] PREDATOR: Registration of 3D Point Clouds with Low Overlap.
Stars: ✭ 106 (-49.28%)
Mutual labels:  attention-mechanism, transformer
X Transformers
A simple but complete full-attention transformer with a set of promising experimental features from various papers
Stars: ✭ 211 (+0.96%)
Pytorch Gat
My implementation of the original GAT paper (Veličković et al.). I've additionally included the playground.py file for visualizing the Cora dataset, GAT embeddings, an attention mechanism, and entropy histograms. I've supported both Cora (transductive) and PPI (inductive) examples!
Stars: ✭ 908 (+334.45%)
Mutual labels:  attention-mechanism, attention
Sockeye
Sequence-to-sequence framework with a focus on Neural Machine Translation based on Apache MXNet
Stars: ✭ 990 (+373.68%)
Mutual labels:  attention-mechanism, transformer
Daily Neural Network Practice 2
Daily Dose of Neural Network that Everyone Needs
Stars: ✭ 18 (-91.39%)
Perceiver Pytorch
Implementation of Perceiver, General Perception with Iterative Attention, in Pytorch
Stars: ✭ 130 (-37.8%)
Prediction Flow
Deep-Learning based CTR models implemented by PyTorch
Stars: ✭ 138 (-33.97%)
Mutual labels:  attention-mechanism, attention
Mlkit
A simple machine learning framework written in Swift 🤖
Stars: ✭ 144 (-31.1%)
Guided Attention Inference Network
Contains implementation of Guided Attention Inference Network (GAIN) presented in Tell Me Where to Look(CVPR 2018). This repository aims to apply GAIN on fcn8 architecture used for segmentation.
Stars: ✭ 204 (-2.39%)
Mutual labels:  attention-mechanism, attention
Machine Learning Course
💬 Machine Learning Course with Python:
Stars: ✭ 6,798 (+3152.63%)
Simplednn
SimpleDNN is a machine learning lightweight open-source library written in Kotlin designed to support relevant neural network architectures in natural language processing tasks
Stars: ✭ 81 (-61.24%)
Nlp Tutorial
Natural Language Processing Tutorial for Deep Learning Researchers
Stars: ✭ 9,895 (+4634.45%)
Mutual labels:  attention, transformer
Eqtransformer
EQTransformer, a python package for earthquake signal detection and phase picking using AI.
Stars: ✭ 95 (-54.55%)
Mutual labels:  attention-mechanism, transformer
Reformer Pytorch
Reformer, the efficient Transformer, in Pytorch
Stars: ✭ 1,644 (+686.6%)
Conformer
Implementation of the convolutional module from the Conformer paper, for use in Transformers
Stars: ✭ 103 (-50.72%)
Machine learning refined
Notes, examples, and Python demos for the textbook "Machine Learning Refined" (published by Cambridge University Press).
Stars: ✭ 750 (+258.85%)
Medical Transformer
Pytorch Code for "Medical Transformer: Gated Axial-Attention for Medical Image Segmentation"
Stars: ✭ 153 (-26.79%)
Mutual labels:  attention, transformer
Image Caption Generator
A neural network to generate captions for an image using CNN and RNN with BEAM Search.
Stars: ✭ 126 (-39.71%)
Mutual labels:  attention-mechanism, attention
Absa keras
Keras Implementation of Aspect based Sentiment Analysis
Stars: ✭ 126 (-39.71%)
Mutual labels:  attention-mechanism, attention
Awesome Quantum Machine Learning
Here you can get all the Quantum Machine learning Basics, Algorithms ,Study Materials ,Projects and the descriptions of the projects around the web
Stars: ✭ 1,940 (+828.23%)
Transformer In Generating Dialogue
An Implementation of 'Attention is all you need' with Chinese Corpus
Stars: ✭ 121 (-42.11%)
Mutual labels:  attention-mechanism, transformer
Sinkhorn Transformer
Sinkhorn Transformer - Practical implementation of Sparse Sinkhorn Attention
Stars: ✭ 156 (-25.36%)
Mariana
The Cutest Deep Learning Framework which is also a wonderful Declarative Language
Stars: ✭ 151 (-27.75%)
Sightseq
Computer vision tools for fairseq, containing PyTorch implementation of text recognition and object detection
Stars: ✭ 116 (-44.5%)
Mutual labels:  attention, transformer
Slot Attention
Implementation of Slot Attention from GoogleAI
Stars: ✭ 168 (-19.62%)
Nano Neuron
🤖 NanoNeuron is 7 simple JavaScript functions that will give you a feeling of how machines can actually "learn"
Stars: ✭ 2,050 (+880.86%)
Dalle Pytorch
Implementation / replication of DALL-E, OpenAI's Text to Image Transformer, in Pytorch
Stars: ✭ 3,661 (+1651.67%)
Awesome Deep Learning For Chinese
最全的中文版深度学习资源索引,包括论文,慕课,开源框架,数据集等等
Stars: ✭ 174 (-16.75%)
Multimodal Sentiment Analysis
Attention-based multimodal fusion for sentiment analysis
Stars: ✭ 172 (-17.7%)
Mutual labels:  attention-mechanism, attention
Transformers.jl
Julia Implementation of Transformer models
Stars: ✭ 173 (-17.22%)
Mutual labels:  attention, transformer
Graph attention pool
Attention over nodes in Graph Neural Networks using PyTorch (NeurIPS 2019)
Stars: ✭ 186 (-11%)
Mutual labels:  attention-mechanism, attention
Log Anomaly Detector
Log Anomaly Detection - Machine learning to detect abnormal events logs
Stars: ✭ 169 (-19.14%)
Datastories Semeval2017 Task4
Deep-learning model presented in "DataStories at SemEval-2017 Task 4: Deep LSTM with Attention for Message-level and Topic-based Sentiment Analysis".
Stars: ✭ 184 (-11.96%)
Mutual labels:  attention-mechanism, attention
Free Ai Resources
🚀 FREE AI Resources - 🎓 Courses, 👷 Jobs, 📝 Blogs, 🔬 AI Research, and many more - for everyone!
Stars: ✭ 192 (-8.13%)
Hnatt
Train and visualize Hierarchical Attention Networks
Stars: ✭ 192 (-8.13%)
Mutual labels:  attention-mechanism, attention
1-60 of 1536 similar projects