danifojo-2018-repeatrnnComparing Fixed and Adaptive Computation Time for Recurrent Neural Networks
Stars: ✭ 32 (-73.98%)
Pytorch Pos TaggingA tutorial on how to implement models for part-of-speech tagging using PyTorch and TorchText.
Stars: ✭ 96 (-21.95%)
Lstm Human Activity RecognitionHuman Activity Recognition example using TensorFlow on smartphone sensors dataset and an LSTM RNN. Classifying the type of movement amongst six activity categories - Guillaume Chevalier
Stars: ✭ 2,943 (+2292.68%)
tiny-rnnLightweight C++11 library for building deep recurrent neural networks
Stars: ✭ 41 (-66.67%)
Gru Svm[ICMLC 2018] A Neural Network Architecture Combining Gated Recurrent Unit (GRU) and Support Vector Machine (SVM) for Intrusion Detection
Stars: ✭ 76 (-38.21%)
EasyesnPython library for Reservoir Computing using Echo State Networks
Stars: ✭ 93 (-24.39%)
Pytorch Sentiment AnalysisTutorials on getting started with PyTorch and TorchText for sentiment analysis.
Stars: ✭ 3,209 (+2508.94%)
Theano Kaldi RnnTHEANO-KALDI-RNNs is a project implementing various Recurrent Neural Networks (RNNs) for RNN-HMM speech recognition. The Theano Code is coupled with the Kaldi decoder.
Stars: ✭ 31 (-74.8%)
SpeakerDiarization RNN CNN LSTMSpeaker Diarization is the problem of separating speakers in an audio. There could be any number of speakers and final result should state when speaker starts and ends. In this project, we analyze given audio file with 2 channels and 2 speakers (on separate channels).
Stars: ✭ 56 (-54.47%)
Linear Attention Recurrent Neural NetworkA recurrent attention module consisting of an LSTM cell which can query its own past cell states by the means of windowed multi-head attention. The formulas are derived from the BN-LSTM and the Transformer Network. The LARNN cell with attention can be easily used inside a loop on the cell state, just like any other RNN. (LARNN)
Stars: ✭ 119 (-3.25%)
Rnn ctcRecurrent Neural Network and Long Short Term Memory (LSTM) with Connectionist Temporal Classification implemented in Theano. Includes a Toy training example.
Stars: ✭ 220 (+78.86%)
Pytorch Kaldipytorch-kaldi is a project for developing state-of-the-art DNN/RNN hybrid speech recognition systems. The DNN part is managed by pytorch, while feature extraction, label computation, and decoding are performed with the kaldi toolkit.
Stars: ✭ 2,097 (+1604.88%)
sequence-rnn-pySequence analyzing using Recurrent Neural Networks (RNN) based on Keras
Stars: ✭ 28 (-77.24%)
Human-Activity-RecognitionHuman activity recognition using TensorFlow on smartphone sensors dataset and an LSTM RNN. Classifying the type of movement amongst six categories (WALKING, WALKING_UPSTAIRS, WALKING_DOWNSTAIRS, SITTING, STANDING, LAYING).
Stars: ✭ 16 (-86.99%)
Codegan[Deprecated] Source Code Generation using Sequence Generative Adversarial Networks
Stars: ✭ 73 (-40.65%)
VariationalNeuralAnnealingA variational implementation of classical and quantum annealing using recurrent neural networks for the purpose of solving optimization problems.
Stars: ✭ 21 (-82.93%)
RnnsharpRNNSharp is a toolkit of deep recurrent neural network which is widely used for many different kinds of tasks, such as sequence labeling, sequence-to-sequence and so on. It's written by C# language and based on .NET framework 4.6 or above versions. RNNSharp supports many different types of networks, such as forward and bi-directional network, sequence-to-sequence network, and different types of layers, such as LSTM, Softmax, sampled Softmax and others.
Stars: ✭ 277 (+125.2%)
RmdlRMDL: Random Multimodel Deep Learning for Classification
Stars: ✭ 375 (+204.88%)
ACTAlternative approach for Adaptive Computation Time in TensorFlow
Stars: ✭ 16 (-86.99%)
DeepseqslamThe Official Deep Learning Framework for Route-based Place Recognition
Stars: ✭ 49 (-60.16%)
sgrnnTensorflow implementation of Synthetic Gradient for RNN (LSTM)
Stars: ✭ 40 (-67.48%)
IseebetteriSeeBetter: Spatio-Temporal Video Super Resolution using Recurrent-Generative Back-Projection Networks | Python3 | PyTorch | GANs | CNNs | ResNets | RNNs | Published in Springer Journal of Computational Visual Media, September 2020, Tsinghua University Press
Stars: ✭ 202 (+64.23%)
Mad TwinnetThe code for the MaD TwinNet. Demo page:
Stars: ✭ 99 (-19.51%)
SleepeegnetSleepEEGNet: Automated Sleep Stage Scoring with Sequence to Sequence Deep Learning Approach
Stars: ✭ 89 (-27.64%)
Malware ClassificationTowards Building an Intelligent Anti-Malware System: A Deep Learning Approach using Support Vector Machine for Malware Classification
Stars: ✭ 88 (-28.46%)
Text predictorChar-level RNN LSTM text generator📄.
Stars: ✭ 99 (-19.51%)
Lstm chemImplementation of the paper - Generative Recurrent Networks for De Novo Drug Design.
Stars: ✭ 87 (-29.27%)
Chemgan ChallengeCode for the paper: Benhenda, M. 2017. ChemGAN challenge for drug discovery: can AI reproduce natural chemical diversity? arXiv preprint arXiv:1708.08227.
Stars: ✭ 98 (-20.33%)
Bit RnnQuantize weights and activations in Recurrent Neural Networks.
Stars: ✭ 86 (-30.08%)
TnnBiologically-realistic recurrent convolutional neural networks
Stars: ✭ 83 (-32.52%)
Skiprnn 2017 TelecombcnSkip RNN: Learning to Skip State Updates in Recurrent Neural Networks (ICLR 2018)
Stars: ✭ 114 (-7.32%)
Lstms.pthPyTorch implementations of LSTM Variants (Dropout + Layer Norm)
Stars: ✭ 111 (-9.76%)
Pytorch EsnAn Echo State Network module for PyTorch.
Stars: ✭ 98 (-20.33%)
CaptcharecognitionEnd-to-end variable length Captcha recognition using CNN+RNN+Attention/CTC (pytorch implementation). 端到端的不定长验证码识别
Stars: ✭ 97 (-21.14%)
SimplednnSimpleDNN is a machine learning lightweight open-source library written in Kotlin designed to support relevant neural network architectures in natural language processing tasks
Stars: ✭ 81 (-34.15%)
Mnist ClassificationPytorch、Scikit-learn实现多种分类方法,包括逻辑回归(Logistic Regression)、多层感知机(MLP)、支持向量机(SVM)、K近邻(KNN)、CNN、RNN,极简代码适合新手小白入门,附英文实验报告(ACM模板)
Stars: ✭ 109 (-11.38%)
EmnistA project designed to explore CNN and the effectiveness of RCNN on classifying the EMNIST dataset.
Stars: ✭ 81 (-34.15%)
Ai Reading MaterialsSome of the ML and DL related reading materials, research papers that I've read
Stars: ✭ 79 (-35.77%)
CodeECG Classification
Stars: ✭ 78 (-36.59%)