PAM[TPAMI 2020] Parallax Attention for Unsupervised Stereo Correspondence Learning
Stars: ✭ 62 (-74.9%)
DrlnDensely Residual Laplacian Super-resolution, IEEE Pattern Analysis and Machine Intelligence (TPAMI), 2020
Stars: ✭ 120 (-51.42%)
Pan[Params: Only 272K!!!] Efficient Image Super-Resolution Using Pixel Attention, in ECCV Workshop, 2020.
Stars: ✭ 151 (-38.87%)
QuantumForestFast Differentiable Forest lib with the advantages of both decision trees and neural networks
Stars: ✭ 63 (-74.49%)
HCFlowOfficial PyTorch code for Hierarchical Conditional Flow: A Unified Framework for Image Super-Resolution and Image Rescaling (HCFlow, ICCV2021)
Stars: ✭ 140 (-43.32%)
pynmta simple and complete pytorch implementation of neural machine translation system
Stars: ✭ 13 (-94.74%)
linformerImplementation of Linformer for Pytorch
Stars: ✭ 119 (-51.82%)
SentimentAnalysisSentiment Analysis: Deep Bi-LSTM+attention model
Stars: ✭ 32 (-87.04%)
srzooSRZoo: An integrated repository for super-resolution using deep learning
Stars: ✭ 135 (-45.34%)
SMSR[CVPR 2021] Exploring Sparsity in Image Super-Resolution for Efficient Inference
Stars: ✭ 205 (-17%)
Transformer-in-TransformerAn Implementation of Transformer in Transformer in TensorFlow for image classification, attention inside local patches
Stars: ✭ 40 (-83.81%)
TMNetThe official pytorch implemention of the CVPR paper "Temporal Modulation Network for Controllable Space-Time Video Super-Resolution".
Stars: ✭ 77 (-68.83%)
nuwa-pytorchImplementation of NÜWA, state of the art attention network for text to video synthesis, in Pytorch
Stars: ✭ 347 (+40.49%)
RAMSOfficial TensorFlow code for paper "Multi-Image Super Resolution of Remotely Sensed Images Using Residual Attention Deep Neural Networks".
Stars: ✭ 55 (-77.73%)
Video-Cap🎬 Video Captioning: ICCV '15 paper implementation
Stars: ✭ 44 (-82.19%)
tensorrt-examplesTensorRT Examples (TensorRT, Jetson Nano, Python, C++)
Stars: ✭ 31 (-87.45%)
enformer-pytorchImplementation of Enformer, Deepmind's attention network for predicting gene expression, in Pytorch
Stars: ✭ 146 (-40.89%)
SwinIRSwinIR: Image Restoration Using Swin Transformer (official repository)
Stars: ✭ 1,260 (+410.12%)
Real-ESRGAN-colabA Real-ESRGAN model trained on a custom dataset
Stars: ✭ 18 (-92.71%)
deepsumDeepSUM: Deep neural network for Super-resolution of Unregistered Multitemporal images (ESA PROBA-V challenge)
Stars: ✭ 39 (-84.21%)
srganPytorch implementation of "Photo-Realistic Single Image Super-Resolution Using a Generative Adversarial Network"
Stars: ✭ 39 (-84.21%)
Patient2VecPatient2Vec: A Personalized Interpretable Deep Representation of the Longitudinal Electronic Health Record
Stars: ✭ 85 (-65.59%)
Super resolution SurveyA survey of recent application of deep learning on super-resolution tasks
Stars: ✭ 32 (-87.04%)
ttslearnttslearn: Library for Pythonで学ぶ音声合成 (Text-to-speech with Python)
Stars: ✭ 158 (-36.03%)
SelfAttentiveImplementation of A Structured Self-attentive Sentence Embedding
Stars: ✭ 107 (-56.68%)
DLSSDeep Learning Super Sampling with Deep Convolutional Generative Adversarial Networks.
Stars: ✭ 88 (-64.37%)
Image-CaptionUsing LSTM or Transformer to solve Image Captioning in Pytorch
Stars: ✭ 36 (-85.43%)
SESRSESR: Single Image Super Resolution with Recursive Squeeze and Excitation Networks
Stars: ✭ 26 (-89.47%)
ntua-slp-semeval2018Deep-learning models of NTUA-SLP team submitted in SemEval 2018 tasks 1, 2 and 3.
Stars: ✭ 79 (-68.02%)
esrganEnhanced SRGAN. Champion PIRM Challenge on Perceptual Super-Resolution
Stars: ✭ 48 (-80.57%)
Attention一些不同的Attention机制代码
Stars: ✭ 17 (-93.12%)
srVAEVAE with RealNVP prior and Super-Resolution VAE in PyTorch. Code release for https://arxiv.org/abs/2006.05218.
Stars: ✭ 56 (-77.33%)
FragmentVCAny-to-any voice conversion by end-to-end extracting and fusing fine-grained voice fragments with attention
Stars: ✭ 134 (-45.75%)
co-attentionPytorch implementation of "Dynamic Coattention Networks For Question Answering"
Stars: ✭ 54 (-78.14%)
nystrom-attentionImplementation of Nyström Self-attention, from the paper Nyströmformer
Stars: ✭ 83 (-66.4%)
NTUA-slp-nlp💻Speech and Natural Language Processing (SLP & NLP) Lab Assignments for ECE NTUA
Stars: ✭ 19 (-92.31%)
NanoJ-FluidicsManual, source-code and binaries for the NanoJ-Fluidics project
Stars: ✭ 47 (-80.97%)
Trident-Dehazing-NetworkNTIRE 2020 NonHomogeneous Dehazing Challenge (CVPR Workshop 2020) 1st Solution.
Stars: ✭ 42 (-83%)
transganformerImplementation of TransGanFormer, an all-attention GAN that combines the finding from the recent GanFormer and TransGan paper
Stars: ✭ 137 (-44.53%)
LightFieldReconstructionHigh-Dimensional Dense Residual Convolutional Neural Network for Light Field Reconstruction
Stars: ✭ 50 (-79.76%)
CF-NetOfficial repository of "Deep Coupled Feedback Network for Joint Exposure Fusion and Image Super-Resolution"
Stars: ✭ 55 (-77.73%)
galerkin-transformer[NeurIPS 2021] Galerkin Transformer: a linear attention without softmax
Stars: ✭ 111 (-55.06%)
keras-deep-learningVarious implementations and projects on CNN, RNN, LSTM, GAN, etc
Stars: ✭ 22 (-91.09%)
MoChA-pytorchPyTorch Implementation of "Monotonic Chunkwise Attention" (ICLR 2018)
Stars: ✭ 65 (-73.68%)
AoA-pytorchA Pytorch implementation of Attention on Attention module (both self and guided variants), for Visual Question Answering
Stars: ✭ 33 (-86.64%)
ADL2019Applied Deep Learning (2019 Spring) @ NTU
Stars: ✭ 20 (-91.9%)
MIRNet-KerasKeras Implementation of MIRNet - SoTA in Image Denoising, Super Resolution and Image Enhancement - CVPR 2020
Stars: ✭ 21 (-91.5%)
GSOCRepository for Google Summer of Code 2019 https://summerofcode.withgoogle.com/projects/#4662790671826944
Stars: ✭ 61 (-75.3%)
traiNNertraiNNer: Deep learning framework for image and video super-resolution, restoration and image-to-image translation, for training and testing.
Stars: ✭ 130 (-47.37%)
RCAN-tfTensorFlow code for ECCV 2018 paper "Image Super-Resolution Using Very Deep Residual Channel Attention Networks"
Stars: ✭ 25 (-89.88%)