Customer-Feedback-AnalysisMulti Class Text (Feedback) Classification using CNN, GRU Network and pre trained Word2Vec embedding, word embeddings on TensorFlow.
Stars: ✭ 18 (-87.92%)
RganRecurrent (conditional) generative adversarial networks for generating real-valued time series data.
Stars: ✭ 480 (+222.15%)
Overlappredator[CVPR 2021, Oral] PREDATOR: Registration of 3D Point Clouds with Low Overlap.
Stars: ✭ 106 (-28.86%)
solar-forecasting-RNNMulti-time-horizon solar forecasting using recurrent neural network
Stars: ✭ 29 (-80.54%)
Srnnsliced-rnn
Stars: ✭ 462 (+210.07%)
Multi-task-Conditional-Attention-NetworksA prototype version of our submitted paper: Conversion Prediction Using Multi-task Conditional Attention Networks to Support the Creation of Effective Ad Creatives.
Stars: ✭ 21 (-85.91%)
stanford-cs231n-assignments-2020This repository contains my solutions to the assignments for Stanford's CS231n "Convolutional Neural Networks for Visual Recognition" (Spring 2020).
Stars: ✭ 84 (-43.62%)
long-short-transformerImplementation of Long-Short Transformer, combining local and global inductive biases for attention over long sequences, in Pytorch
Stars: ✭ 103 (-30.87%)
Skip Thoughts.torchPorting of Skip-Thoughts pretrained models from Theano to PyTorch & Torch7
Stars: ✭ 146 (-2.01%)
visdialVisual Dialog: Light-weight Transformer for Many Inputs (ECCV 2020)
Stars: ✭ 27 (-81.88%)
Transformer TtsA Pytorch Implementation of "Neural Speech Synthesis with Transformer Network"
Stars: ✭ 418 (+180.54%)
Codegan[Deprecated] Source Code Generation using Sequence Generative Adversarial Networks
Stars: ✭ 73 (-51.01%)
Pytorch Original TransformerMy implementation of the original transformer model (Vaswani et al.). I've additionally included the playground.py file for visualizing otherwise seemingly hard concepts. Currently included IWSLT pretrained models.
Stars: ✭ 411 (+175.84%)
Lstms.pthPyTorch implementations of LSTM Variants (Dropout + Layer Norm)
Stars: ✭ 111 (-25.5%)
CrabNetPredict materials properties using only the composition information!
Stars: ✭ 57 (-61.74%)
Se3 Transformer PytorchImplementation of SE3-Transformers for Equivariant Self-Attention, in Pytorch. This specific repository is geared towards integration with eventual Alphafold2 replication.
Stars: ✭ 73 (-51.01%)
cnn-rnn-bitcoinReusable CNN and RNN model doing time series binary classification
Stars: ✭ 28 (-81.21%)
Wavetorch 🌊 Numerically solving and backpropagating through the wave equation
Stars: ✭ 387 (+159.73%)
ArrayLSTMGPU/CPU (CUDA) Implementation of "Recurrent Memory Array Structures", Simple RNN, LSTM, Array LSTM..
Stars: ✭ 21 (-85.91%)
Image Caption GeneratorA neural network to generate captions for an image using CNN and RNN with BEAM Search.
Stars: ✭ 126 (-15.44%)
stylegan-pokemonGenerating Pokemon cards using a mixture of StyleGAN and RNN to create beautiful & vibrant cards ready for battle!
Stars: ✭ 47 (-68.46%)
RmdlRMDL: Random Multimodel Deep Learning for Classification
Stars: ✭ 375 (+151.68%)
Machine-Translation-Hindi-to-english-Machine translation is the task of converting one language to other. Unlike the traditional phrase-based translation system which consists of many small sub-components that are tuned separately, neural machine translation attempts to build and train a single, large neural network that reads a sentence and outputs a correct translation.
Stars: ✭ 19 (-87.25%)
h-transformer-1dImplementation of H-Transformer-1D, Hierarchical Attention for Sequence Learning
Stars: ✭ 121 (-18.79%)
Mnist ClassificationPytorch、Scikit-learn实现多种分类方法,包括逻辑回归(Logistic Regression)、多层感知机(MLP)、支持向量机(SVM)、K近邻(KNN)、CNN、RNN,极简代码适合新手小白入门,附英文实验报告(ACM模板)
Stars: ✭ 109 (-26.85%)
DCAN[AAAI 2020] Code release for "Domain Conditioned Adaptation Network" https://arxiv.org/abs/2005.06717
Stars: ✭ 27 (-81.88%)
SimgnnA PyTorch implementation of "SimGNN: A Neural Network Approach to Fast Graph Similarity Computation" (WSDM 2019).
Stars: ✭ 351 (+135.57%)
IndRNN pytorchIndependently Recurrent Neural Networks (IndRNN) implemented in pytorch.
Stars: ✭ 112 (-24.83%)
tf-ran-cellRecurrent Additive Networks for Tensorflow
Stars: ✭ 16 (-89.26%)
Fast PytorchPytorch Tutorial, Pytorch with Google Colab, Pytorch Implementations: CNN, RNN, DCGAN, Transfer Learning, Chatbot, Pytorch Sample Codes
Stars: ✭ 346 (+132.21%)
axial-attentionImplementation of Axial attention - attending to multi-dimensional data efficiently
Stars: ✭ 245 (+64.43%)
Attribute Aware Attention[ACM MM 2018] Attribute-Aware Attention Model for Fine-grained Representation Learning
Stars: ✭ 143 (-4.03%)
TransformerA TensorFlow Implementation of the Transformer: Attention Is All You Need
Stars: ✭ 3,646 (+2346.98%)
Rep-CounterAI Exercise Rep Counter based on Google's Human Pose Estimation Library (Posenet)
Stars: ✭ 47 (-68.46%)
Plasma PythonPPPL deep learning disruption prediction package
Stars: ✭ 65 (-56.38%)
QuantumForestFast Differentiable Forest lib with the advantages of both decision trees and neural networks
Stars: ✭ 63 (-57.72%)
Keras GatKeras implementation of the graph attention networks (GAT) by Veličković et al. (2017; https://arxiv.org/abs/1710.10903)
Stars: ✭ 334 (+124.16%)
RnnoiseRecurrent neural network for audio noise reduction
Stars: ✭ 2,266 (+1420.81%)
Vae SeqVariational Auto-Encoders in a Sequential Setting.
Stars: ✭ 145 (-2.68%)
SleepeegnetSleepEEGNet: Automated Sleep Stage Scoring with Sequence to Sequence Deep Learning Approach
Stars: ✭ 89 (-40.27%)