All Projects → ixaxaar → Pytorch Sublstm

ixaxaar / Pytorch Sublstm

Licence: mit
subLSTMs for pytorch from Cortical microcircuits as gated-recurrent neural networks

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to Pytorch Sublstm

Vad
Voice activity detection (VAD) toolkit including DNN, bDNN, LSTM and ACAM based VAD. We also provide our directly recorded dataset.
Stars: ✭ 622 (+4046.67%)
Mutual labels:  lstm
Python autocomplete
A simple neural network for python autocompletion
Stars: ✭ 779 (+5093.33%)
Mutual labels:  lstm
Chainer Rnn Ner
Named Entity Recognition with RNN, implemented by Chainer
Stars: ✭ 19 (+26.67%)
Mutual labels:  lstm
Conv Emotion
This repo contains implementation of different architectures for emotion recognition in conversations.
Stars: ✭ 646 (+4206.67%)
Mutual labels:  lstm
Getting Things Done With Pytorch
Jupyter Notebook tutorials on solving real-world problems with Machine Learning & Deep Learning using PyTorch. Topics: Face detection with Detectron 2, Time Series anomaly detection with LSTM Autoencoders, Object Detection with YOLO v5, Build your first Neural Network, Time Series forecasting for Coronavirus daily cases, Sentiment Analysis with BERT.
Stars: ✭ 738 (+4820%)
Mutual labels:  lstm
Tensorflow Tutorial
Some interesting TensorFlow tutorials for beginners.
Stars: ✭ 893 (+5853.33%)
Mutual labels:  lstm
Multi Class Text Classification Cnn Rnn
Classify Kaggle San Francisco Crime Description into 39 classes. Build the model with CNN, RNN (GRU and LSTM) and Word Embeddings on Tensorflow.
Stars: ✭ 570 (+3700%)
Mutual labels:  lstm
Spago
Self-contained Machine Learning and Natural Language Processing library in Go
Stars: ✭ 854 (+5593.33%)
Mutual labels:  lstm
Seq2seq Chatbot
Chatbot in 200 lines of code using TensorLayer
Stars: ✭ 777 (+5080%)
Mutual labels:  lstm
Multitask Learning Protein Prediction
Multitask learning: protein secondary structure prediction, b-values prediction and solvent-accessibility prediction
Stars: ✭ 19 (+26.67%)
Mutual labels:  lstm
Text Classification
Implementation of papers for text classification task on DBpedia
Stars: ✭ 682 (+4446.67%)
Mutual labels:  lstm
Lstm Char Cnn Tensorflow
in progress
Stars: ✭ 737 (+4813.33%)
Mutual labels:  lstm
Music2dance
Generating Dance steps for given music with deep learning
Stars: ✭ 17 (+13.33%)
Mutual labels:  lstm
Ad examples
A collection of anomaly detection methods (iid/point-based, graph and time series) including active learning for anomaly detection/discovery, bayesian rule-mining, description for diversity/explanation/interpretability. Analysis of incorporating label feedback with ensemble and tree-based detectors. Includes adversarial attacks with Graph Convolutional Network.
Stars: ✭ 641 (+4173.33%)
Mutual labels:  lstm
Deep Music Genre Classification
🎵 Using Deep Learning to Categorize Music as Time Progresses Through Spectrogram Analysis
Stars: ✭ 23 (+53.33%)
Mutual labels:  lstm
Telemanom
A framework for using LSTMs to detect anomalies in multivariate time series data. Includes spacecraft anomaly data and experiments from the Mars Science Laboratory and SMAP missions.
Stars: ✭ 589 (+3826.67%)
Mutual labels:  lstm
Deep Learning Time Series
List of papers, code and experiments using deep learning for time series forecasting
Stars: ✭ 796 (+5206.67%)
Mutual labels:  lstm
Neural Image Captioning
Implementation of Neural Image Captioning model using Keras with Theano backend
Stars: ✭ 12 (-20%)
Mutual labels:  lstm
Re Verb
speaker diarization system using an LSTM
Stars: ✭ 27 (+80%)
Mutual labels:  lstm
Tensorflow Onnx
Convert TensorFlow models to ONNX
Stars: ✭ 900 (+5900%)
Mutual labels:  lstm

subtractive LSTM (subLSTM), for Pytorch

Build Status PyPI version

This is an implementation of subLSTM described in the paper Cortical microcircuits as gated-recurrent neural networks, Rui Ponte Costa et al.

Install

pip install pytorch-sublstm

Usage

Parameters:

Following are the constructor parameters:

Argument Default Description
input_size None Size of the input vectors
hidden_size None Size of hidden units
num_layers 1 Number of layers in the network
bias True Bias
batch_first False Whether data is fed batch first
dropout 0 Dropout between layers in the network
bidirectional False If the network is bidirectional

Example usage:

nn Interface

import torch
from torch.autograd import Variable
from subLSTM.nn import SubLSTM

hidden_size = 20
input_size = 10
seq_len = 5
batch_size = 7
hidden = None

input = Variable(torch.randn(batch_size, seq_len, input_size))

rnn = SubLSTM(input_size, hidden_size, num_layers=2, bias=True, batch_first=True)

# forward pass
output, hidden = rnn(input, hidden)

Cell Interface

import torch
from torch.autograd import Variable
from subLSTM.nn import SubLSTMCell

hidden_size = 20
input_size = 10
seq_len = 5
batch_size = 7
hidden = None

hx = Variable(torch.randn(batch_size, hidden_size))
cx = Variable(torch.randn(batch_size, hidden_size))

input = Variable(torch.randn(batch_size, input_size))

cell = SubLSTMCell(input_size, hidden_size, bias=True)
(hx, cx) = cell(input, (hx, cx))

Tasks:

A language modeling task is included here. Refer to its README for more info.

Attributions:

A lot of the code is recycled from pytorch

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].