All Projects → imatge-upc → danifojo-2018-repeatrnn

imatge-upc / danifojo-2018-repeatrnn

Licence: MIT license
Comparing Fixed and Adaptive Computation Time for Recurrent Neural Networks

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to danifojo-2018-repeatrnn

Codegan
[Deprecated] Source Code Generation using Sequence Generative Adversarial Networks
Stars: ✭ 73 (+128.13%)
Mutual labels:  recurrent-neural-networks, rnn
Pytorch Learners Tutorial
PyTorch tutorial for learners
Stars: ✭ 97 (+203.13%)
Mutual labels:  recurrent-neural-networks, rnn
Gru Svm
[ICMLC 2018] A Neural Network Architecture Combining Gated Recurrent Unit (GRU) and Support Vector Machine (SVM) for Intrusion Detection
Stars: ✭ 76 (+137.5%)
Mutual labels:  recurrent-neural-networks, rnn
Predicting Myers Briggs Type Indicator With Recurrent Neural Networks
Stars: ✭ 43 (+34.38%)
Mutual labels:  recurrent-neural-networks, rnn
Deep Learning With Python
Deep learning codes and projects using Python
Stars: ✭ 195 (+509.38%)
Mutual labels:  recurrent-neural-networks, rnn
Deepseqslam
The Official Deep Learning Framework for Route-based Place Recognition
Stars: ✭ 49 (+53.13%)
Mutual labels:  recurrent-neural-networks, rnn
Pytorch Pos Tagging
A tutorial on how to implement models for part-of-speech tagging using PyTorch and TorchText.
Stars: ✭ 96 (+200%)
Mutual labels:  recurrent-neural-networks, rnn
Rnnsharp
RNNSharp is a toolkit of deep recurrent neural network which is widely used for many different kinds of tasks, such as sequence labeling, sequence-to-sequence and so on. It's written by C# language and based on .NET framework 4.6 or above versions. RNNSharp supports many different types of networks, such as forward and bi-directional network, sequence-to-sequence network, and different types of layers, such as LSTM, Softmax, sampled Softmax and others.
Stars: ✭ 277 (+765.63%)
Mutual labels:  recurrent-neural-networks, rnn
Pytorch Kaldi
pytorch-kaldi is a project for developing state-of-the-art DNN/RNN hybrid speech recognition systems. The DNN part is managed by pytorch, while feature extraction, label computation, and decoding are performed with the kaldi toolkit.
Stars: ✭ 2,097 (+6453.13%)
Mutual labels:  recurrent-neural-networks, rnn
Rnn From Scratch
Use tensorflow's tf.scan to build vanilla, GRU and LSTM RNNs
Stars: ✭ 123 (+284.38%)
Mutual labels:  recurrent-neural-networks, rnn
Theano Kaldi Rnn
THEANO-KALDI-RNNs is a project implementing various Recurrent Neural Networks (RNNs) for RNN-HMM speech recognition. The Theano Code is coupled with the Kaldi decoder.
Stars: ✭ 31 (-3.12%)
Mutual labels:  recurrent-neural-networks, rnn
Rnn ctc
Recurrent Neural Network and Long Short Term Memory (LSTM) with Connectionist Temporal Classification implemented in Theano. Includes a Toy training example.
Stars: ✭ 220 (+587.5%)
Mutual labels:  recurrent-neural-networks, rnn
Tensorflow Char Rnn
Char-RNN implemented using TensorFlow.
Stars: ✭ 429 (+1240.63%)
Mutual labels:  recurrent-neural-networks, rnn
Bitcoin Price Prediction Using Lstm
Bitcoin price Prediction ( Time Series ) using LSTM Recurrent neural network
Stars: ✭ 67 (+109.38%)
Mutual labels:  recurrent-neural-networks, rnn
Rmdl
RMDL: Random Multimodel Deep Learning for Classification
Stars: ✭ 375 (+1071.88%)
Mutual labels:  recurrent-neural-networks, rnn
Easyesn
Python library for Reservoir Computing using Echo State Networks
Stars: ✭ 93 (+190.63%)
Mutual labels:  recurrent-neural-networks, rnn
EdgarAllanPoetry
Computer-generated poetry
Stars: ✭ 22 (-31.25%)
Mutual labels:  recurrent-neural-networks, rnn
Lstm Human Activity Recognition
Human Activity Recognition example using TensorFlow on smartphone sensors dataset and an LSTM RNN. Classifying the type of movement amongst six activity categories - Guillaume Chevalier
Stars: ✭ 2,943 (+9096.88%)
Mutual labels:  recurrent-neural-networks, rnn
Linear Attention Recurrent Neural Network
A recurrent attention module consisting of an LSTM cell which can query its own past cell states by the means of windowed multi-head attention. The formulas are derived from the BN-LSTM and the Transformer Network. The LARNN cell with attention can be easily used inside a loop on the cell state, just like any other RNN. (LARNN)
Stars: ✭ 119 (+271.88%)
Mutual labels:  recurrent-neural-networks, rnn
Iseebetter
iSeeBetter: Spatio-Temporal Video Super Resolution using Recurrent-Generative Back-Projection Networks | Python3 | PyTorch | GANs | CNNs | ResNets | RNNs | Published in Springer Journal of Computational Visual Media, September 2020, Tsinghua University Press
Stars: ✭ 202 (+531.25%)
Mutual labels:  recurrent-neural-networks, rnn

Adaptive Computation Time for Recurrent Neural Networks in PyTorch and Tensorflow

We recommend using the TensorFlow implementation, as it is much faster and it was tested further.

To use the PyTorch implementation switch to the "PyTorch" branch.

To install TensorFlow or PyTorch, follow the online instructions: TensorFlow, PyTorch.

To run one of the tasks available tasks with ACT just run:

python addition.py

or

python parity.py

To see available options, run:

python addition.py -h

If you want to apply ACT to your own RNN cell, call ACTCell with your RNN as input. You can see examples in the code for our tasks.

To test the new baseline, run (only in TensorFlow):

python addition-repeat.py

or

python parity-repeat.py
Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].