All Projects → Transformer → Similar Projects or Alternatives

336 Open source projects that are alternatives of or similar to Transformer

Sockeye
Sequence-to-sequence framework with a focus on Neural Machine Translation based on Apache MXNet
Stars: ✭ 990 (+265.31%)
Nmt Keras
Neural Machine Translation with Keras
Stars: ✭ 501 (+84.87%)
Pytorch Original Transformer
My implementation of the original transformer model (Vaswani et al.). I've additionally included the playground.py file for visualizing otherwise seemingly hard concepts. Currently included IWSLT pretrained models.
Stars: ✭ 411 (+51.66%)
SequenceToSequence
A seq2seq with attention dialogue/MT model implemented by TensorFlow.
Stars: ✭ 11 (-95.94%)
Machine Translation
Stars: ✭ 51 (-81.18%)
Linear Attention Recurrent Neural Network
A recurrent attention module consisting of an LSTM cell which can query its own past cell states by the means of windowed multi-head attention. The formulas are derived from the BN-LSTM and the Transformer Network. The LARNN cell with attention can be easily used inside a loop on the cell state, just like any other RNN. (LARNN)
Stars: ✭ 119 (-56.09%)
Transformer
A TensorFlow Implementation of the Transformer: Attention Is All You Need
Stars: ✭ 3,646 (+1245.39%)
Transformers without tears
Transformers without Tears: Improving the Normalization of Self-Attention
Stars: ✭ 80 (-70.48%)
Witwicky
Witwicky: An implementation of Transformer in PyTorch.
Stars: ✭ 21 (-92.25%)
Attention Mechanisms
Implementations for a family of attention mechanisms, suitable for all kinds of natural language processing tasks and compatible with TensorFlow 2.0 and Keras.
Stars: ✭ 203 (-25.09%)
Machine-Translation-Hindi-to-english-
Machine translation is the task of converting one language to other. Unlike the traditional phrase-based translation system which consists of many small sub-components that are tuned separately, neural machine translation attempts to build and train a single, large neural network that reads a sentence and outputs a correct translation.
Stars: ✭ 19 (-92.99%)
pynmt
a simple and complete pytorch implementation of neural machine translation system
Stars: ✭ 13 (-95.2%)
Mutual labels:  attention-mechanism
PAM
[TPAMI 2020] Parallax Attention for Unsupervised Stereo Correspondence Learning
Stars: ✭ 62 (-77.12%)
Mutual labels:  attention-mechanism
attention-mechanism-keras
attention mechanism in keras, like Dense and RNN...
Stars: ✭ 19 (-92.99%)
Mutual labels:  attention-mechanism
FragmentVC
Any-to-any voice conversion by end-to-end extracting and fusing fine-grained voice fragments with attention
Stars: ✭ 134 (-50.55%)
Mutual labels:  attention-mechanism
transganformer
Implementation of TransGanFormer, an all-attention GAN that combines the finding from the recent GanFormer and TransGan paper
Stars: ✭ 137 (-49.45%)
Mutual labels:  attention-mechanism
Attention
一些不同的Attention机制代码
Stars: ✭ 17 (-93.73%)
Mutual labels:  attention-mechanism
nystrom-attention
Implementation of Nyström Self-attention, from the paper Nyströmformer
Stars: ✭ 83 (-69.37%)
Mutual labels:  attention-mechanism
Patient2Vec
Patient2Vec: A Personalized Interpretable Deep Representation of the Longitudinal Electronic Health Record
Stars: ✭ 85 (-68.63%)
Mutual labels:  attention-mechanism
keras attention
🔖 An Attention Layer in Keras
Stars: ✭ 43 (-84.13%)
Mutual labels:  attention-mechanism
AoA-pytorch
A Pytorch implementation of Attention on Attention module (both self and guided variants), for Visual Question Answering
Stars: ✭ 33 (-87.82%)
Mutual labels:  attention-mechanism
automatic-personality-prediction
[AAAI 2020] Modeling Personality with Attentive Networks and Contextual Embeddings
Stars: ✭ 43 (-84.13%)
Mutual labels:  attention-mechanism
banglanmt
This repository contains the code and data of the paper titled "Not Low-Resource Anymore: Aligner Ensembling, Batch Filtering, and New Datasets for Bengali-English Machine Translation" published in Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP 2020), November 16 - November 20, 2020.
Stars: ✭ 91 (-66.42%)
Mutual labels:  machine-translation
ADL2019
Applied Deep Learning (2019 Spring) @ NTU
Stars: ✭ 20 (-92.62%)
Mutual labels:  attention-mechanism
MoChA-pytorch
PyTorch Implementation of "Monotonic Chunkwise Attention" (ICLR 2018)
Stars: ✭ 65 (-76.01%)
Mutual labels:  attention-mechanism
OverlapPredator
[CVPR 2021, Oral] PREDATOR: Registration of 3D Point Clouds with Low Overlap.
Stars: ✭ 293 (+8.12%)
Mutual labels:  attention-mechanism
QuantumForest
Fast Differentiable Forest lib with the advantages of both decision trees and neural networks
Stars: ✭ 63 (-76.75%)
Mutual labels:  attention-mechanism
Attention-Visualization
Visualization for simple attention and Google's multi-head attention.
Stars: ✭ 54 (-80.07%)
Mutual labels:  machine-translation
Video-Cap
🎬 Video Captioning: ICCV '15 paper implementation
Stars: ✭ 44 (-83.76%)
Mutual labels:  attention-mechanism
linformer
Implementation of Linformer for Pytorch
Stars: ✭ 119 (-56.09%)
Mutual labels:  attention-mechanism
SentimentAnalysis
Sentiment Analysis: Deep Bi-LSTM+attention model
Stars: ✭ 32 (-88.19%)
Mutual labels:  attention-mechanism
transformer-pytorch
A PyTorch implementation of Transformer in "Attention is All You Need"
Stars: ✭ 77 (-71.59%)
Mutual labels:  machine-translation
enformer-pytorch
Implementation of Enformer, Deepmind's attention network for predicting gene expression, in Pytorch
Stars: ✭ 146 (-46.13%)
Mutual labels:  attention-mechanism
Writing-editing-Network
Code for Paper Abstract Writing through Editing Mechanism
Stars: ✭ 72 (-73.43%)
Mutual labels:  attention-mechanism
NTUA-slp-nlp
💻Speech and Natural Language Processing (SLP & NLP) Lab Assignments for ECE NTUA
Stars: ✭ 19 (-92.99%)
Mutual labels:  attention-mechanism
co-attention
Pytorch implementation of "Dynamic Coattention Networks For Question Answering"
Stars: ✭ 54 (-80.07%)
Mutual labels:  attention-mechanism
keras-deep-learning
Various implementations and projects on CNN, RNN, LSTM, GAN, etc
Stars: ✭ 22 (-91.88%)
Mutual labels:  attention-mechanism
BangalASR
Transformer based Bangla Speech Recognition
Stars: ✭ 20 (-92.62%)
Mutual labels:  attention-is-all-you-need
ntua-slp-semeval2018
Deep-learning models of NTUA-SLP team submitted in SemEval 2018 tasks 1, 2 and 3.
Stars: ✭ 79 (-70.85%)
Mutual labels:  attention-mechanism
Natural-Language-Processing
Contains various architectures and novel paper implementations for Natural Language Processing tasks like Sequence Modelling and Neural Machine Translation.
Stars: ✭ 48 (-82.29%)
Mutual labels:  machine-translation
DAF3D
Deep Attentive Features for Prostate Segmentation in 3D Transrectal Ultrasound
Stars: ✭ 60 (-77.86%)
Mutual labels:  attention-mechanism
Mirnet
Official repository for "Learning Enriched Features for Real Image Restoration and Enhancement" (ECCV 2020). SOTA results for image denoising, super-resolution, and image enhancement.
Stars: ✭ 247 (-8.86%)
Mutual labels:  attention-mechanism
CompareModels TRECQA
Compare six baseline deep learning models on TrecQA
Stars: ✭ 61 (-77.49%)
Mutual labels:  attention-mechanism
NLP Toolkit
Library of state-of-the-art models (PyTorch) for NLP tasks
Stars: ✭ 92 (-66.05%)
Mutual labels:  machine-translation
visualization
a collection of visualization function
Stars: ✭ 189 (-30.26%)
Mutual labels:  attention-mechanism
dcsp segmentation
No description or website provided.
Stars: ✭ 34 (-87.45%)
Mutual labels:  attention-mechanism
vista-net
Code for the paper "VistaNet: Visual Aspect Attention Network for Multimodal Sentiment Analysis", AAAI'19
Stars: ✭ 67 (-75.28%)
Mutual labels:  attention-mechanism
ttslearn
ttslearn: Library for Pythonで学ぶ音声合成 (Text-to-speech with Python)
Stars: ✭ 158 (-41.7%)
Mutual labels:  attention-mechanism
SANET
Arbitrary Style Transfer with Style-Attentional Networks
Stars: ✭ 105 (-61.25%)
Mutual labels:  attention-mechanism
halonet-pytorch
Implementation of the 😇 Attention layer from the paper, Scaling Local Self-Attention For Parameter Efficient Visual Backbones
Stars: ✭ 181 (-33.21%)
Mutual labels:  attention-mechanism
apertium-html-tools
Web application providing a fully localised interface for text/website/document translation, analysis and generation powered by Apertium.
Stars: ✭ 36 (-86.72%)
Mutual labels:  machine-translation
egfr-att
Drug effect prediction using neural network
Stars: ✭ 17 (-93.73%)
Mutual labels:  attention-mechanism
ilmulti
Tooling to play around with multilingual machine translation for Indian Languages.
Stars: ✭ 19 (-92.99%)
Mutual labels:  machine-translation
galerkin-transformer
[NeurIPS 2021] Galerkin Transformer: a linear attention without softmax
Stars: ✭ 111 (-59.04%)
Mutual labels:  attention-mechanism
Retinal-Disease-Diagnosis-With-Residual-Attention-Networks
Using Residual Attention Networks to diagnose retinal diseases in medical images
Stars: ✭ 14 (-94.83%)
Mutual labels:  attention-mechanism
Transformer-in-Transformer
An Implementation of Transformer in Transformer in TensorFlow for image classification, attention inside local patches
Stars: ✭ 40 (-85.24%)
Mutual labels:  attention-mechanism
Multigrid-Neural-Architectures
Multigrid Neural Architecture
Stars: ✭ 28 (-89.67%)
Mutual labels:  attention-mechanism
Video-guided-Machine-Translation
Starter code for the VMT task and challenge
Stars: ✭ 45 (-83.39%)
Mutual labels:  machine-translation
SelfAttentive
Implementation of A Structured Self-attentive Sentence Embedding
Stars: ✭ 107 (-60.52%)
Mutual labels:  attention-mechanism
3HAN
An original implementation of "3HAN: A Deep Neural Network for Fake News Detection" (ICONIP 2017)
Stars: ✭ 29 (-89.3%)
Mutual labels:  attention-mechanism
1-60 of 336 similar projects