ixaxaar / Pytorch Sublstm
Licence: mit
subLSTMs for pytorch from Cortical microcircuits as gated-recurrent neural networks
Stars: ✭ 15
Programming Languages
python
139335 projects - #7 most used programming language
Labels
Projects that are alternatives of or similar to Pytorch Sublstm
Vad
Voice activity detection (VAD) toolkit including DNN, bDNN, LSTM and ACAM based VAD. We also provide our directly recorded dataset.
Stars: ✭ 622 (+4046.67%)
Mutual labels: lstm
Python autocomplete
A simple neural network for python autocompletion
Stars: ✭ 779 (+5093.33%)
Mutual labels: lstm
Chainer Rnn Ner
Named Entity Recognition with RNN, implemented by Chainer
Stars: ✭ 19 (+26.67%)
Mutual labels: lstm
Conv Emotion
This repo contains implementation of different architectures for emotion recognition in conversations.
Stars: ✭ 646 (+4206.67%)
Mutual labels: lstm
Getting Things Done With Pytorch
Jupyter Notebook tutorials on solving real-world problems with Machine Learning & Deep Learning using PyTorch. Topics: Face detection with Detectron 2, Time Series anomaly detection with LSTM Autoencoders, Object Detection with YOLO v5, Build your first Neural Network, Time Series forecasting for Coronavirus daily cases, Sentiment Analysis with BERT.
Stars: ✭ 738 (+4820%)
Mutual labels: lstm
Tensorflow Tutorial
Some interesting TensorFlow tutorials for beginners.
Stars: ✭ 893 (+5853.33%)
Mutual labels: lstm
Multi Class Text Classification Cnn Rnn
Classify Kaggle San Francisco Crime Description into 39 classes. Build the model with CNN, RNN (GRU and LSTM) and Word Embeddings on Tensorflow.
Stars: ✭ 570 (+3700%)
Mutual labels: lstm
Spago
Self-contained Machine Learning and Natural Language Processing library in Go
Stars: ✭ 854 (+5593.33%)
Mutual labels: lstm
Seq2seq Chatbot
Chatbot in 200 lines of code using TensorLayer
Stars: ✭ 777 (+5080%)
Mutual labels: lstm
Multitask Learning Protein Prediction
Multitask learning: protein secondary structure prediction, b-values prediction and solvent-accessibility prediction
Stars: ✭ 19 (+26.67%)
Mutual labels: lstm
Text Classification
Implementation of papers for text classification task on DBpedia
Stars: ✭ 682 (+4446.67%)
Mutual labels: lstm
Music2dance
Generating Dance steps for given music with deep learning
Stars: ✭ 17 (+13.33%)
Mutual labels: lstm
Ad examples
A collection of anomaly detection methods (iid/point-based, graph and time series) including active learning for anomaly detection/discovery, bayesian rule-mining, description for diversity/explanation/interpretability. Analysis of incorporating label feedback with ensemble and tree-based detectors. Includes adversarial attacks with Graph Convolutional Network.
Stars: ✭ 641 (+4173.33%)
Mutual labels: lstm
Deep Music Genre Classification
🎵 Using Deep Learning to Categorize Music as Time Progresses Through Spectrogram Analysis
Stars: ✭ 23 (+53.33%)
Mutual labels: lstm
Telemanom
A framework for using LSTMs to detect anomalies in multivariate time series data. Includes spacecraft anomaly data and experiments from the Mars Science Laboratory and SMAP missions.
Stars: ✭ 589 (+3826.67%)
Mutual labels: lstm
Deep Learning Time Series
List of papers, code and experiments using deep learning for time series forecasting
Stars: ✭ 796 (+5206.67%)
Mutual labels: lstm
Neural Image Captioning
Implementation of Neural Image Captioning model using Keras with Theano backend
Stars: ✭ 12 (-20%)
Mutual labels: lstm
subtractive LSTM (subLSTM), for Pytorch
This is an implementation of subLSTM described in the paper Cortical microcircuits as gated-recurrent neural networks, Rui Ponte Costa et al.
Install
pip install pytorch-sublstm
Usage
Parameters:
Following are the constructor parameters:
Argument | Default | Description |
---|---|---|
input_size | None |
Size of the input vectors |
hidden_size | None |
Size of hidden units |
num_layers | 1 |
Number of layers in the network |
bias | True |
Bias |
batch_first | False |
Whether data is fed batch first |
dropout | 0 |
Dropout between layers in the network |
bidirectional | False |
If the network is bidirectional |
Example usage:
nn Interface
import torch
from torch.autograd import Variable
from subLSTM.nn import SubLSTM
hidden_size = 20
input_size = 10
seq_len = 5
batch_size = 7
hidden = None
input = Variable(torch.randn(batch_size, seq_len, input_size))
rnn = SubLSTM(input_size, hidden_size, num_layers=2, bias=True, batch_first=True)
# forward pass
output, hidden = rnn(input, hidden)
Cell Interface
import torch
from torch.autograd import Variable
from subLSTM.nn import SubLSTMCell
hidden_size = 20
input_size = 10
seq_len = 5
batch_size = 7
hidden = None
hx = Variable(torch.randn(batch_size, hidden_size))
cx = Variable(torch.randn(batch_size, hidden_size))
input = Variable(torch.randn(batch_size, input_size))
cell = SubLSTMCell(input_size, hidden_size, bias=True)
(hx, cx) = cell(input, (hx, cx))
Tasks:
A language modeling task is included here. Refer to its README for more info.
Attributions:
A lot of the code is recycled from pytorch
Note that the project description data, including the texts, logos, images, and/or trademarks,
for each open source project belongs to its rightful owner.
If you wish to add or remove any projects, please contact us at [email protected].