All Projects → titu1994 → Keras Sru

titu1994 / Keras Sru

Implementation of Simple Recurrent Unit in Keras

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to Keras Sru

Tensorflow Lstm Sin
TensorFlow 1.3 experiment with LSTM (and GRU) RNNs for sine prediction
Stars: ✭ 52 (-31.58%)
Mutual labels:  recurrent-neural-networks
Dogtorch
Who Let The Dogs Out? Modeling Dog Behavior From Visual Data https://arxiv.org/pdf/1803.10827.pdf
Stars: ✭ 66 (-13.16%)
Mutual labels:  recurrent-neural-networks
Keras Tcn
Keras Temporal Convolutional Network.
Stars: ✭ 1,177 (+1448.68%)
Mutual labels:  recurrent-neural-networks
Simplehtr
Handwritten Text Recognition (HTR) system implemented with TensorFlow.
Stars: ✭ 1,072 (+1310.53%)
Mutual labels:  recurrent-neural-networks
Rcnn Relation Extraction
Tensorflow Implementation of Recurrent Convolutional Neural Network for Relation Extraction
Stars: ✭ 64 (-15.79%)
Mutual labels:  recurrent-neural-networks
Bitcoin Price Prediction Using Lstm
Bitcoin price Prediction ( Time Series ) using LSTM Recurrent neural network
Stars: ✭ 67 (-11.84%)
Mutual labels:  recurrent-neural-networks
Tensorflow Cnn Time Series
Feeding images of time series to Conv Nets! (Tensorflow + Keras)
Stars: ✭ 49 (-35.53%)
Mutual labels:  recurrent-neural-networks
Codegan
[Deprecated] Source Code Generation using Sequence Generative Adversarial Networks
Stars: ✭ 73 (-3.95%)
Mutual labels:  recurrent-neural-networks
Chicksexer
A Python package for gender classification.
Stars: ✭ 64 (-15.79%)
Mutual labels:  recurrent-neural-networks
3d Reconstruction With Neural Networks
3D reconstruction with neural networks using Tensorflow. See link for Video (https://www.youtube.com/watch?v=iI6ZMST8Ri0)
Stars: ✭ 71 (-6.58%)
Mutual labels:  recurrent-neural-networks
Sentiment Analysis Nltk Ml Lstm
Sentiment Analysis on the First Republic Party debate in 2016 based on Python,NLTK and ML.
Stars: ✭ 61 (-19.74%)
Mutual labels:  recurrent-neural-networks
Predrnn Pytorch
Official implementation for NIPS'17 paper: PredRNN: Recurrent Neural Networks for Predictive Learning Using Spatiotemporal LSTMs.
Stars: ✭ 59 (-22.37%)
Mutual labels:  recurrent-neural-networks
Torch Ac
Recurrent and multi-process PyTorch implementation of deep reinforcement Actor-Critic algorithms A2C and PPO
Stars: ✭ 70 (-7.89%)
Mutual labels:  recurrent-neural-networks
Image Captioning
Image Captioning: Implementing the Neural Image Caption Generator with python
Stars: ✭ 52 (-31.58%)
Mutual labels:  recurrent-neural-networks
Rnn Trajmodel
The source of the IJCAI2017 paper "Modeling Trajectory with Recurrent Neural Networks"
Stars: ✭ 72 (-5.26%)
Mutual labels:  recurrent-neural-networks
Deepseqslam
The Official Deep Learning Framework for Route-based Place Recognition
Stars: ✭ 49 (-35.53%)
Mutual labels:  recurrent-neural-networks
Deep Plant
Deep-Plant: Plant Classification with CNN/RNN. It consists of CAFFE/Tensorflow implementation of our PR-17, TIP-18 (HGO-CNN & PlantStructNet) and MalayaKew dataset.
Stars: ✭ 66 (-13.16%)
Mutual labels:  recurrent-neural-networks
Qrnn
Quasi-recurrent Neural Networks for Keras
Stars: ✭ 74 (-2.63%)
Mutual labels:  recurrent-neural-networks
Recurrent Environment Simulators
Deepmind Recurrent Environment Simulators paper implementation in tensorflow
Stars: ✭ 73 (-3.95%)
Mutual labels:  recurrent-neural-networks
Lstm Ctc Ocr
using rnn (lstm or gru) and ctc to convert line image into text, based on torch7 and warp-ctc
Stars: ✭ 70 (-7.89%)
Mutual labels:  recurrent-neural-networks

Keras Simple Recurrent Unit (SRU)

Implementation of Simple Recurrent Unit in Keras. Paper - Training RNNs as Fast as CNNs

This is a naive implementation with some speed gains over the generic LSTM cells, however its speed is not yet 10x that of cuDNN LSTMs

Issues

  • [x] Fix the need to unroll the SRU to get it to work correctly

  • [x] -Input dim must exactly match the number of LSTM cells for now. Still working out how to overcome this problem.-

No longer a problem to have different input dimension than output dimension.

  • [x] Performance of a single SRU layer is slightly lower (about 0.5% on average over 5 runs) compared to 1 layer LSTM (at least on IMDB, with batch size of 32). Haven't tried staking them yet, but this may improve performance.

Performance degrades substantially with larger batch sizes (about 6-7% on average over 5 runs) compared to 1 layer LSTM with batch size of 128. However, a multi layer SRU (I've tried with 3 layers), while a bit slower than a 1 layer LSTM, gets around the same score on batch size of 32 or 128.

Seems the solution to this is to stack several SRUs together. The authors recommend stacks of 4 SRU layers.

  • [ ] Speed gains aren't that impressive at small batch size. At batch size of 32, SRU takes around 32-34 seconds. LSTM takes around 60-70 seconds. Thats just 50% reduction in speed, not the 5-10x that was discussed in the paper.

However, once batch size is increased to 128, SRU takes just 7 seconds per epoch compared to LSTM 22 seconds. For comparison, CNNs take 3-4 seconds per epoch.

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].