All Projects → kracwarlock → Action Recognition Visual Attention

kracwarlock / Action Recognition Visual Attention

Action recognition using soft attention based deep recurrent neural networks

Projects that are alternatives of or similar to Action Recognition Visual Attention

Sparse Structured Attention
Sparse and structured neural attention mechanisms
Stars: ✭ 198 (-43.43%)
Mutual labels:  deep-neural-networks, deeplearning, attention-mechanism
Magnet
Deep Learning Projects that Build Themselves
Stars: ✭ 351 (+0.29%)
Mutual labels:  jupyter-notebook, deep-neural-networks, deeplearning
Ludwig
Data-centric declarative deep learning framework
Stars: ✭ 8,018 (+2190.86%)
Mutual labels:  deep-neural-networks, deeplearning, deep
Pytorch Original Transformer
My implementation of the original transformer model (Vaswani et al.). I've additionally included the playground.py file for visualizing otherwise seemingly hard concepts. Currently included IWSLT pretrained models.
Stars: ✭ 411 (+17.43%)
Mutual labels:  jupyter-notebook, deeplearning, attention-mechanism
Attention is all you need
Transformer of "Attention Is All You Need" (Vaswani et al. 2017) by Chainer.
Stars: ✭ 303 (-13.43%)
Mutual labels:  jupyter-notebook, deep-neural-networks, attention-mechanism
Learnopencv
Learn OpenCV : C++ and Python Examples
Stars: ✭ 15,385 (+4295.71%)
Mutual labels:  jupyter-notebook, deep-neural-networks, deeplearning
Best ai paper 2020
A curated list of the latest breakthroughs in AI by release date with a clear video explanation, link to a more in-depth article, and code
Stars: ✭ 2,140 (+511.43%)
Mutual labels:  paper, deep-neural-networks, deeplearning
Deep Embedded Memory Networks
https://arxiv.org/abs/1707.00836
Stars: ✭ 19 (-94.57%)
Mutual labels:  jupyter-notebook, paper, deep-neural-networks
Concise Ipython Notebooks For Deep Learning
Ipython Notebooks for solving problems like classification, segmentation, generation using latest Deep learning algorithms on different publicly available text and image data-sets.
Stars: ✭ 23 (-93.43%)
Mutual labels:  jupyter-notebook, deep-neural-networks, deeplearning
Servenet
Service Classification based on Service Description
Stars: ✭ 21 (-94%)
Mutual labels:  jupyter-notebook, deep-neural-networks, deeplearning
Triplet Attention
Official PyTorch Implementation for "Rotate to Attend: Convolutional Triplet Attention Module." [WACV 2021]
Stars: ✭ 222 (-36.57%)
Mutual labels:  jupyter-notebook, paper, attention-mechanism
ghiaseddin
Author's implementation of the paper "Deep Relative Attributes" (ACCV 2016)
Stars: ✭ 41 (-88.29%)
Mutual labels:  paper, deeplearning
T81 558 deep learning
Washington University (in St. Louis) Course T81-558: Applications of Deep Neural Networks
Stars: ✭ 4,152 (+1086.29%)
Mutual labels:  jupyter-notebook, deeplearning
SentimentAnalysis
Sentiment Analysis: Deep Bi-LSTM+attention model
Stars: ✭ 32 (-90.86%)
Mutual labels:  deeplearning, attention-mechanism
Dlpython course
Примеры для курса "Программирование глубоких нейронных сетей на Python"
Stars: ✭ 266 (-24%)
Mutual labels:  jupyter-notebook, deep-neural-networks
efficient-attention
An implementation of the efficient attention module.
Stars: ✭ 191 (-45.43%)
Mutual labels:  paper, attention-mechanism
Da Rnn
📃 **Unofficial** PyTorch Implementation of DA-RNN (arXiv:1704.02971)
Stars: ✭ 256 (-26.86%)
Mutual labels:  jupyter-notebook, attention-mechanism
Deeplearning.ai Assignments
Stars: ✭ 268 (-23.43%)
Mutual labels:  jupyter-notebook, deeplearning
Dab
Data Augmentation by Backtranslation (DAB) ヽ( •_-)ᕗ
Stars: ✭ 294 (-16%)
Mutual labels:  jupyter-notebook, deep-neural-networks
Cascaded Fcn
Source code for the MICCAI 2016 Paper "Automatic Liver and Lesion Segmentation in CT Using Cascaded Fully Convolutional NeuralNetworks and 3D Conditional Random Fields"
Stars: ✭ 296 (-15.43%)
Mutual labels:  jupyter-notebook, deep-neural-networks

Action Recognition using Visual Attention

We propose a soft attention based model for the task of action recognition in videos. We use multi-layered Recurrent Neural Networks (RNNs) with Long-Short Term Memory (LSTM) units which are deep both spatially and temporally. Our model learns to focus selectively on parts of the video frames and classifies videos after taking a few glimpses. The model essentially learns which parts in the frames are relevant for the task at hand and attaches higher importance to them. We evaluate the model on UCF-11 (YouTube Action), HMDB-51 and Hollywood2 datasets and analyze how the model focuses its attention depending on the scene and the action being performed.

Dependencies

Input data format

This is provided in util/README.md

Reference

If you use this code as part of any published research, please acknowledge the following papers:

"Action Recognition using Visual Attention."
Shikhar Sharma, Ryan Kiros, Ruslan Salakhutdinov. arXiv

@article{sharma2015attention,
    title={Action Recognition using Visual Attention},
    author={Sharma, Shikhar and Kiros, Ryan and Salakhutdinov, Ruslan},
    journal={arXiv preprint arXiv:1511.04119},
    year={2015}
} 

"Show, Attend and Tell: Neural Image Caption Generation with Visual Attention."
Kelvin Xu, Jimmy Ba, Ryan Kiros, Kyunghyun Cho, Aaron Courville, Ruslan Salakhutdinov, Richard Zemel, Yoshua Bengio. To appear ICML (2015)

@article{Xu2015show,
    title={Show, Attend and Tell: Neural Image Caption Generation with Visual Attention},
    author={Xu, Kelvin and Ba, Jimmy and Kiros, Ryan and Cho, Kyunghyun and Courville, Aaron and Salakhutdinov, Ruslan and Zemel, Richard and Bengio, Yoshua},
    journal={arXiv preprint arXiv:1502.03044},
    year={2015}
}

License

This repsoitory is released under a revised (3-clause) BSD License. It is the implementation for our paper Action Recognition using Visual Attention. The repository uses some code from the project arctic-caption which is originally the implementation for the paper Show, Attend and Tell: Neural Image Caption Generation with Visual Attention and is also licensed under a revised (3-clause) BSD License.

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].