All Projects → transganformer → Similar Projects or Alternatives

400 Open source projects that are alternatives of or similar to transganformer

ntua-slp-semeval2018
Deep-learning models of NTUA-SLP team submitted in SemEval 2018 tasks 1, 2 and 3.
Stars: ✭ 79 (-42.34%)
Mutual labels:  attention-mechanism
knowledge-neurons
A library for finding knowledge neurons in pretrained transformer models.
Stars: ✭ 72 (-47.45%)
Mutual labels:  transformers
shoe-design-using-generative-adversarial-networks
No description or website provided.
Stars: ✭ 18 (-86.86%)
CompareModels TRECQA
Compare six baseline deep learning models on TrecQA
Stars: ✭ 61 (-55.47%)
Mutual labels:  attention-mechanism
DAF3D
Deep Attentive Features for Prostate Segmentation in 3D Transrectal Ultrasound
Stars: ✭ 60 (-56.2%)
Mutual labels:  attention-mechanism
trapper
State-of-the-art NLP through transformer models in a modular design and consistent APIs.
Stars: ✭ 28 (-79.56%)
Mutual labels:  transformers
BERT-NER
Using pre-trained BERT models for Chinese and English NER with 🤗Transformers
Stars: ✭ 114 (-16.79%)
Mutual labels:  transformers
TorchBlocks
A PyTorch-based toolkit for natural language processing
Stars: ✭ 85 (-37.96%)
Mutual labels:  transformers
remixer-pytorch
Implementation of the Remixer Block from the Remixer paper, in Pytorch
Stars: ✭ 37 (-72.99%)
Mutual labels:  transformers
mSRGAN-A-GAN-for-single-image-super-resolution-on-high-content-screening-microscopy-images.
Generative Adversarial Network for single image super-resolution in high content screening microscopy images
Stars: ✭ 52 (-62.04%)
optimum
🏎️ Accelerate training and inference of 🤗 Transformers with easy to use hardware optimization tools
Stars: ✭ 567 (+313.87%)
Mutual labels:  transformers
SelfAttentive
Implementation of A Structured Self-attentive Sentence Embedding
Stars: ✭ 107 (-21.9%)
Mutual labels:  attention-mechanism
iPerceive
Applying Common-Sense Reasoning to Multi-Modal Dense Video Captioning and Video Question Answering | Python3 | PyTorch | CNNs | Causality | Reasoning | LSTMs | Transformers | Multi-Head Self Attention | Published in IEEE Winter Conference on Applications of Computer Vision (WACV) 2021
Stars: ✭ 52 (-62.04%)
Mutual labels:  transformers
vista-net
Code for the paper "VistaNet: Visual Aspect Attention Network for Multimodal Sentiment Analysis", AAAI'19
Stars: ✭ 67 (-51.09%)
Mutual labels:  attention-mechanism
SANET
Arbitrary Style Transfer with Style-Attentional Networks
Stars: ✭ 105 (-23.36%)
Mutual labels:  attention-mechanism
halonet-pytorch
Implementation of the 😇 Attention layer from the paper, Scaling Local Self-Attention For Parameter Efficient Visual Backbones
Stars: ✭ 181 (+32.12%)
Mutual labels:  attention-mechanism
TermiNetwork
🌏 A zero-dependency networking solution for building modern and secure iOS, watchOS, macOS and tvOS applications.
Stars: ✭ 80 (-41.61%)
Mutual labels:  transformers
TransCenter
This is the official implementation of TransCenter. The code and pretrained models are now available here: https://gitlab.inria.fr/yixu/TransCenter_official.
Stars: ✭ 82 (-40.15%)
Mutual labels:  transformers
PAM
[TPAMI 2020] Parallax Attention for Unsupervised Stereo Correspondence Learning
Stars: ✭ 62 (-54.74%)
Mutual labels:  attention-mechanism
Semi-Supervised-Learning-GAN
Semi-supervised Learning GAN
Stars: ✭ 72 (-47.45%)
pynmt
a simple and complete pytorch implementation of neural machine translation system
Stars: ✭ 13 (-90.51%)
Mutual labels:  attention-mechanism
stylegan-v
[CVPR 2022] StyleGAN-V: A Continuous Video Generator with the Price, Image Quality and Perks of StyleGAN2
Stars: ✭ 136 (-0.73%)
spark-transformers
Spark-Transformers: Library for exporting Apache Spark MLLIB models to use them in any Java application with no other dependencies.
Stars: ✭ 39 (-71.53%)
Mutual labels:  transformers
course-content-dl
NMA deep learning course
Stars: ✭ 537 (+291.97%)
Mutual labels:  transformers
OpenDialog
An Open-Source Package for Chinese Open-domain Conversational Chatbot (中文闲聊对话系统,一键部署微信闲聊机器人)
Stars: ✭ 94 (-31.39%)
Mutual labels:  transformers
deepfrog
An NLP-suite powered by deep learning
Stars: ✭ 16 (-88.32%)
Mutual labels:  transformers
KoELECTRA-Pipeline
Transformers Pipeline with KoELECTRA
Stars: ✭ 37 (-72.99%)
Mutual labels:  transformers
pytorch-vit
An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale
Stars: ✭ 250 (+82.48%)
Mutual labels:  transformers
BangalASR
Transformer based Bangla Speech Recognition
Stars: ✭ 20 (-85.4%)
Mutual labels:  transformers
domain-attention
codes for paper "Domain Attention Model for Multi-Domain Sentiment Classification"
Stars: ✭ 22 (-83.94%)
Mutual labels:  attention-mechanism
attention-mechanism-keras
attention mechanism in keras, like Dense and RNN...
Stars: ✭ 19 (-86.13%)
Mutual labels:  attention-mechanism
abcnn pytorch
Implementation of ABCNN(Attention-Based Convolutional Neural Network) on Pytorch
Stars: ✭ 35 (-74.45%)
Mutual labels:  attention-mechanism
C-Tran
General Multi-label Image Classification with Transformers
Stars: ✭ 106 (-22.63%)
Mutual labels:  transformers
converse
Conversational text Analysis using various NLP techniques
Stars: ✭ 147 (+7.3%)
Mutual labels:  transformers
SentimentAnalysis
Sentiment Analysis: Deep Bi-LSTM+attention model
Stars: ✭ 32 (-76.64%)
Mutual labels:  attention-mechanism
geometry-free-view-synthesis
Is a geometric model required to synthesize novel views from a single image?
Stars: ✭ 265 (+93.43%)
Mutual labels:  transformers
simple transformers
Simple transformer implementations that I can understand
Stars: ✭ 18 (-86.86%)
Mutual labels:  transformers
modules
The official repository for our paper "Are Neural Nets Modular? Inspecting Functional Modularity Through Differentiable Weight Masks". We develop a method for analyzing emerging functional modularity in neural networks based on differentiable weight masks and use it to point out important issues in current-day neural networks.
Stars: ✭ 25 (-81.75%)
Mutual labels:  transformers
FragmentVC
Any-to-any voice conversion by end-to-end extracting and fusing fine-grained voice fragments with attention
Stars: ✭ 134 (-2.19%)
Mutual labels:  attention-mechanism
extkeras
Playground for implementing custom layers and other components compatible with keras, with the purpose to learn the framework better and perhaps in future offer some utils for others.
Stars: ✭ 18 (-86.86%)
Mutual labels:  attention-mechanism
text2text
Text2Text: Cross-lingual natural language processing and generation toolkit
Stars: ✭ 188 (+37.23%)
Mutual labels:  transformers
enformer-pytorch
Implementation of Enformer, Deepmind's attention network for predicting gene expression, in Pytorch
Stars: ✭ 146 (+6.57%)
Mutual labels:  attention-mechanism
hamnet
PyTorch implementation of AAAI 2021 paper: A Hybrid Attention Mechanism for Weakly-Supervised Temporal Action Localization
Stars: ✭ 30 (-78.1%)
Mutual labels:  attention-mechanism
HugsVision
HugsVision is a easy to use huggingface wrapper for state-of-the-art computer vision
Stars: ✭ 154 (+12.41%)
Mutual labels:  transformers
datastories-semeval2017-task6
Deep-learning model presented in "DataStories at SemEval-2017 Task 6: Siamese LSTM with Attention for Humorous Text Comparison".
Stars: ✭ 20 (-85.4%)
Mutual labels:  attention-mechanism
text2keywords
Trained T5 and T5-large model for creating keywords from text
Stars: ✭ 53 (-61.31%)
Mutual labels:  transformers
xpandas
Universal 1d/2d data containers with Transformers functionality for data analysis.
Stars: ✭ 25 (-81.75%)
Mutual labels:  transformers
serverless-transformers-on-aws-lambda
Deploy transformers serverless on AWS Lambda
Stars: ✭ 100 (-27.01%)
Mutual labels:  transformers
NTUA-slp-nlp
💻Speech and Natural Language Processing (SLP & NLP) Lab Assignments for ECE NTUA
Stars: ✭ 19 (-86.13%)
Mutual labels:  attention-mechanism
WellcomeML
Repository for Machine Learning utils at the Wellcome Trust
Stars: ✭ 31 (-77.37%)
Mutual labels:  transformers
bert-squeeze
🛠️ Tools for Transformers compression using PyTorch Lightning ⚡
Stars: ✭ 56 (-59.12%)
Mutual labels:  transformers
gan-vae-pretrained-pytorch
Pretrained GANs + VAEs + classifiers for MNIST/CIFAR in pytorch.
Stars: ✭ 134 (-2.19%)
fusion gan
Codes for the paper 'Learning to Fuse Music Genres with Generative Adversarial Dual Learning' ICDM 17
Stars: ✭ 18 (-86.86%)
text2class
Multi-class text categorization using state-of-the-art pre-trained contextualized language models, e.g. BERT
Stars: ✭ 15 (-89.05%)
Mutual labels:  transformers
eve-bot
EVE bot, a customer service chatbot to enhance virtual engagement for Twitter Apple Support
Stars: ✭ 31 (-77.37%)
Mutual labels:  transformers
lightning-transformers
Flexible components pairing 🤗 Transformers with Pytorch Lightning
Stars: ✭ 551 (+302.19%)
Mutual labels:  transformers
dodrio
Exploring attention weights in transformer-based models with linguistic knowledge.
Stars: ✭ 233 (+70.07%)
Mutual labels:  attention-mechanism
GAN-Project-2018
GAN in Tensorflow to be run via Linux command line
Stars: ✭ 21 (-84.67%)
Patient2Vec
Patient2Vec: A Personalized Interpretable Deep Representation of the Longitudinal Electronic Health Record
Stars: ✭ 85 (-37.96%)
Mutual labels:  attention-mechanism
61-120 of 400 similar projects