All Projects → Caoyichao → STCNet

Caoyichao / STCNet

Licence: other
STCNet: Spatio-Temporal Cross Network for Industrial Smoke Detection

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to STCNet

Temporally Language Grounding
A Pytorch implemention for some state-of-the-art models for" Temporally Language Grounding in Untrimmed Videos"
Stars: ✭ 73 (+151.72%)
Mutual labels:  video-understanding
Awesome Activity Prediction
Paper list of activity prediction and related area
Stars: ✭ 147 (+406.9%)
Mutual labels:  video-understanding
Awesome Grounding
awesome grounding: A curated list of research papers in visual grounding
Stars: ✭ 247 (+751.72%)
Mutual labels:  video-understanding
Temporal Segment Networks
Code & Models for Temporal Segment Networks (TSN) in ECCV 2016
Stars: ✭ 1,287 (+4337.93%)
Mutual labels:  video-understanding
Multiverse
Dataset, code and model for the CVPR'20 paper "The Garden of Forking Paths: Towards Multi-Future Trajectory Prediction". And for the ECCV'20 SimAug paper.
Stars: ✭ 131 (+351.72%)
Mutual labels:  video-understanding
Youtube 8m
The 2nd place Solution to the Youtube-8M Video Understanding Challenge by Team Monkeytyping (based on tensorflow)
Stars: ✭ 171 (+489.66%)
Mutual labels:  video-understanding
Tsn Pytorch
Temporal Segment Networks (TSN) in PyTorch
Stars: ✭ 895 (+2986.21%)
Mutual labels:  video-understanding
NExT-QA
NExT-QA: Next Phase of Question-Answering to Explaining Temporal Actions (CVPR'21)
Stars: ✭ 50 (+72.41%)
Mutual labels:  video-understanding
Video2tfrecord
Easily convert RGB video data (e.g. .avi) to the TensorFlow tfrecords file format for training e.g. a NN in TensorFlow. This implementation allows to limit the number of frames per video to be stored in the tfrecords.
Stars: ✭ 137 (+372.41%)
Mutual labels:  video-understanding
Paddlevideo
Comprehensive, latest, and deployable video deep learning algorithm, including video recognition, action localization, and temporal action detection tasks. It's a high-performance, light-weight codebase provides practical models for video understanding research and application
Stars: ✭ 218 (+651.72%)
Mutual labels:  video-understanding
Movienet Tools
Tools for movie and video research
Stars: ✭ 113 (+289.66%)
Mutual labels:  video-understanding
Mmaction
An open-source toolbox for action understanding based on PyTorch
Stars: ✭ 1,711 (+5800%)
Mutual labels:  video-understanding
Step
STEP: Spatio-Temporal Progressive Learning for Video Action Detection. CVPR'19 (Oral)
Stars: ✭ 196 (+575.86%)
Mutual labels:  video-understanding
Temporal Shift Module
[ICCV 2019] TSM: Temporal Shift Module for Efficient Video Understanding
Stars: ✭ 1,282 (+4320.69%)
Mutual labels:  video-understanding
SSTDA
[CVPR 2020] Action Segmentation with Joint Self-Supervised Temporal Domain Adaptation (PyTorch)
Stars: ✭ 150 (+417.24%)
Mutual labels:  video-understanding
Tdn
[CVPR 2021] TDN: Temporal Difference Networks for Efficient Action Recognition
Stars: ✭ 72 (+148.28%)
Mutual labels:  video-understanding
Object level visual reasoning
Pytorch Implementation of "Object level Visual Reasoning in Videos", F. Baradel, N. Neverova, C. Wolf, J. Mille, G. Mori , ECCV 2018
Stars: ✭ 163 (+462.07%)
Mutual labels:  video-understanding
MTL-AQA
What and How Well You Performed? A Multitask Learning Approach to Action Quality Assessment [CVPR 2019]
Stars: ✭ 38 (+31.03%)
Mutual labels:  video-understanding
glimpse clouds
Pytorch implementation of the paper "Glimpse Clouds: Human Activity Recognition from Unstructured Feature Points", F. Baradel, C. Wolf, J. Mille , G.W. Taylor, CVPR 2018
Stars: ✭ 30 (+3.45%)
Mutual labels:  video-understanding
Actionvlad
ActionVLAD for video action classification (CVPR 2017)
Stars: ✭ 217 (+648.28%)
Mutual labels:  video-understanding

STCNet

STCNet: Spatio-Temporal Cross Network for Industrial Smoke Detection

The code will be available soon.

Environment

Python 3.6

Pytorch 1.3+

Experiments

F-Scores for some methods on RISE dataset.

Model S0 S1 S2 S3 S4 S5 Average
Flow-SVM .42 .59 .47 .63 .52 .47 .517
Flow-I3D .55 .58 .51 .68 .65 .50 .578
RGB-SVM .57 .70 .67 .67 .57 .53 .618
RGB-I3D .80 .84 .82 .87 .82 .75 .817
RGB-I3D-ND .76 .79 .81 .86 .76 .68 .777
RGB-I3D-FP .76 .81 .82 .87 .81 .71 .797
RGB-I3D-TSM .81 .84 .82 .87 .80 .74 .813
RGB-I3D-LSTM .80 .84 .82 .85 .83 .74 .813
RGB-I3D-NL .81 .84 .83 .87 .81 .74 .817
RGB-I3D-TC .81 .84 .84 .87 .81 .77 .823
Plain SE-Resnext .83 .82 .84 .85 .78 .83 .826
STCNet(MobileNetv2) .86 .88 .87 .89 .84 .86 .868
STCNet(SE-ResNext) .88 .89 .90 .90 .86 .88 .885

Compare with other methods on RISE dataset. (RTX2080Ti GPU)

Model Backbone Params Flops Latency Throughput Average
RGB-I3D Inception I3D 12.3M 62.7G 30.56ms 32.71vid/s .817
RGB-I3D-TSM Inception I3D 12.3M 62.7G 31.85ms 31.40vid/s .813
RGB-I3D-LSTM Inception I3D 38.0M 62.9G 31.01ms 32.25vid/s .813
RGB-I3D-NL Inception I3D 12.3M 62.7G 30.32ms 32.98vid/s .817
RGB-I3D-TC Inception I3D 12.3M 62.7G 30.41ms 32.88vid/s .823
Plain SE-Resnext SE-ResNeXt-50 26.6M 34.4G 22.10ms 45.25vid/s .826
STCNet (Proposed) Mobilenetv2 3.7M 2.4G 9.12ms 109.7vid/s .868
STCNet (Proposed) SE-ResNeXt-50 27.2M 34.6G 23.49ms 42.57vid/s .885

Visualization

Input RGB frames (the top row) in RISE dataset and corresponding residual frames (the bottom row)

Grad-CAM visualization for spatial and temporal pathway.

GRAD-CAM visualization of Spatial path:

GRAD-CAM visualization of Temporal path:

False positive cases in the testing set.

False negative cases in the testing set.

Each GIF has the same name as the original video. If interested, you can check the corresponding original video in RISE dataset: https://github.com/CMU-CREATE-Lab/deep-smoke-machine

Acknowledgements

We thank Carnegie Mellon University (CMU) and Pennsylvania State University (PSU) for their efforts in environmental protection. We also thank the Big Data Center of Southeast University for providing the facility support on the numerical calculations in this paper.

And this is a good implementation for our method: https://github.com/ChangyWen/STCNet-for-Smoke-Detection

Citation

If you use our code or paper, please cite:

Y. Cao, Q. Tang, X. Lu, F. Li, and J. Cao, “STCNet: Spatio-Temporal Cross Network for Industrial Smoke Detection,” arXiv:2011.04863 [cs], Nov. 2020, Accessed: Nov. 16, 2020. [Online]. Available: http://arxiv.org/abs/2011.04863.

Contact

If you have any question, please feel free to contact me (Yichao Cao, [email protected]). Thanks :-)

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].