All Projects → DengPingFan → Davsod

DengPingFan / Davsod

Shifting More Attention to Video Salient Objection Detection, CVPR 2019 (Best paper finalist & Oral)

Projects that are alternatives of or similar to Davsod

Mastering Python For Finance Second Edition
Mastering Python for Finance – Second Edition, published by Packt
Stars: ✭ 153 (-1.29%)
Mutual labels:  jupyter-notebook
Deep Viz Keras
Implementations of some popular Saliency Maps in Keras
Stars: ✭ 154 (-0.65%)
Mutual labels:  jupyter-notebook
Binderhub
Run your code in the cloud, with technology so advanced, it feels like magic!
Stars: ✭ 2,050 (+1222.58%)
Mutual labels:  jupyter-notebook
Raster Vision Examples
Examples of using Raster Vision on open datasets
Stars: ✭ 154 (-0.65%)
Mutual labels:  jupyter-notebook
Data Science Stack Cookiecutter
🐳📊🤓Cookiecutter template to launch an awesome dockerized Data Science toolstack (incl. Jupyster, Superset, Postgres, Minio, AirFlow & API Star)
Stars: ✭ 153 (-1.29%)
Mutual labels:  jupyter-notebook
Cnnvis Pytorch
visualization of CNN in PyTorch
Stars: ✭ 154 (-0.65%)
Mutual labels:  jupyter-notebook
Self Learning
Books Papers, Courses & more I have to learn soon
Stars: ✭ 146 (-5.81%)
Mutual labels:  jupyter-notebook
Pyportfolioopt
Financial portfolio optimisation in python, including classical efficient frontier, Black-Litterman, Hierarchical Risk Parity
Stars: ✭ 2,502 (+1514.19%)
Mutual labels:  jupyter-notebook
Example Seldon
Example for end-to-end machine learning on Kubernetes using Kubeflow and Seldon Core
Stars: ✭ 154 (-0.65%)
Mutual labels:  jupyter-notebook
Jupyter Server Proxy
Jupyter notebook server extension to proxy web services.
Stars: ✭ 153 (-1.29%)
Mutual labels:  jupyter-notebook
Daguan Classify 2018
2018达观杯长文本分类智能处理挑战赛 18解决方案
Stars: ✭ 154 (-0.65%)
Mutual labels:  jupyter-notebook
Automatic Generation Of Text Summaries
使用两种方法(抽取式Textrank和概要式seq2seq)自动提取文本摘要
Stars: ✭ 155 (+0%)
Mutual labels:  jupyter-notebook
Matplotlib Label Lines
Label line using matplotlib.
Stars: ✭ 154 (-0.65%)
Mutual labels:  jupyter-notebook
Robuststl
Unofficial Implementation of RobustSTL: A Robust Seasonal-Trend Decomposition Algorithm for Long Time Series (AAAI 2019)
Stars: ✭ 154 (-0.65%)
Mutual labels:  jupyter-notebook
Neural Style Transfer
Keras Implementation of Neural Style Transfer from the paper "A Neural Algorithm of Artistic Style" (http://arxiv.org/abs/1508.06576) in Keras 2.0+
Stars: ✭ 2,000 (+1190.32%)
Mutual labels:  jupyter-notebook
Python For Probability Statistics And Machine Learning 2e
Second edition of Springer Book Python for Probability, Statistics, and Machine Learning
Stars: ✭ 154 (-0.65%)
Mutual labels:  jupyter-notebook
Tensorflow Multi Dimensional Lstm
Multi dimensional LSTM as described in Alex Graves' Paper https://arxiv.org/pdf/0705.2011.pdf
Stars: ✭ 154 (-0.65%)
Mutual labels:  jupyter-notebook
Jupyter Vim Binding
Jupyter meets Vim. Vimmer will fall in love.
Stars: ✭ 1,965 (+1167.74%)
Mutual labels:  jupyter-notebook
Deepreinforcementlearning
A replica of the AlphaZero methodology for deep reinforcement learning in Python
Stars: ✭ 1,898 (+1124.52%)
Mutual labels:  jupyter-notebook
Stock Market Prediction Challenge
Following repo is the solution to Stock Market Prediction using Neural Networks and Sentiment Analysis
Stars: ✭ 154 (-0.65%)
Mutual labels:  jupyter-notebook

SSAV (CVPR2019-Oral)

Code for paper in CVPR2019, 'Shifting More Attention to Video Salient Object Detection', Deng-Ping Fan, Wenguan Wang, Ming-Ming Cheng, Jianbing Shen.

Contact: Deng-Ping Fan, Email: [email protected]

Paper with code: https://paperswithcode.com/task/video-salient-object-detection ("[D*F*n], [S**V2*19c*prbe****ark]")

SSAV framework

Abstract The last decade has witnessed a growing interest in video salient object detection (VSOD). However, the research community long-term lacked a well-established VSOD dataset representative of real dynamic scenes with high-quality annotations. To address this issue, we elaborately collected a visual-attention-consistent Densely Annotated VSOD (DAVSOD) dataset, which contains 226 videos with 23,938 frames that cover diverse realistic-scenes, objects, instances and motions. With corresponding real human eye-fixation data, we obtain precise ground-truths. This is the first work that explicitly emphasizes the challenge of saliency shift, i.e., the video salient object(s) may dynamically change. To further contribute the community a complete benchmark, we systematically assess 17 representative VSOD algorithms over seven existing VSOD datasets and our DAVSOD with totally ~84K frames (largest-scale). Utilizing three famous metrics, we then present a comprehensive and insightful performance analysis. Furthermore, we propose a baseline model. It is equipped with a saliencyshift-aware convLSTM, which can efficiently capture video saliency dynamics through learning human attention-shift behavior. Extensive experiments1 open up promising future directions for model development and comparison.

pre-computed saliency maps and Datasets: http://dpfan.net/DAVSOD/.

Blackswan sequence result generated by our SSAV model. More results can be found in our webside http://dpfan.net/DAVSOD/.

Usage

  1. Clone this repo into your computer
git clone https://github.com/DengPingFan/DAVSOD.git
  1. Cd to DAVSOD/mycaffe-convlstm, follow the official instructions to build caffe. We provide our make file Makefile.config in folder DAVSOD/mycaffe-convlstm.

The code has been tested successfully on Ubuntu 16.04 with CUDA 8.0 and OpenCV 3.1.0

  1. Make 'caffe'
make all -j8
  1. Make 'pycaffe'
make pycaffe
  1. Download pretrained caffemodel from my homepage or directly from [baidu pan](Fetch Code: pb0h)/ [google drive] and extract the .zip file under the root directory DAVSOD/model/. If you want to train the model start from scratch, you can download the basemodel from [baidu pan](Fetch Code:0xk4) or [google drive]

  2. Put the test image in DAVSOD/Datasets/ and run generateTestList.py to get the test list. Then run SSAV_test.py to get the saliency maps. The results will be saved in DAVSOD/results/SSAV/.

  3. You can also evaluate the model performance (S-measure[1], E-measure[2], F-measure and MAE) using our one-key matlab code main.m in DAVSOD/EvaluateTool/ directory.

[1]Structure-measure: A New Way to Evaluate the Foregournd Maps, ICCV2017.
[2]Enhanced Alignment Measure for Binary Foreground Map Evaluation, IJCAI2018.

Note that: This version only provide the implicit manner for learning attention-shift. The explicit way to train this model will not be released due to the commercial purposes (Hua Wei, IIAI).


Performance Preview

Quantitative comparisons table4

Quanlitative comparisons figure6

Related Citations (BibTeX)

If you find this useful, please cite the related works as follows: SSAV model/DAVSOD dataset

@InProceedings{Fan_2019_CVPR,
   author = {Fan, Deng-Ping and Wang, Wenguan and Cheng, Ming-Ming and Shen, Jianbing}, 
   title = {Shifting More Attention to Video Salient Object Detection},
   booktitle = {IEEE CVPR},
   year = {2019}
}
@inproceedings{song2018pyramid,
  title={Pyramid dilated deeper ConvLSTM for video salient object detection},
  author={Song, Hongmei and Wang, Wenguan and Zhao, Sanyuan and Shen, Jianbing and Lam, Kin-Man},
  booktitle={Proceedings of the European Conference on Computer Vision},
  pages={715--731},
  year={2018}
}
@InProceedings{Wang_2018_CVPR,
author = {Wang, Wenguan and Shen, Jianbing and Guo, Fang and Cheng, Ming-Ming and Borji, Ali},
title = {Revisiting Video Saliency: A Large-Scale Benchmark and a New Model},
booktitle = {The IEEE Conference on Computer Vision and Pattern Recognition},
year = {2018}
}

Metrics

%E-measure
@inproceedings{Fan2018Enhanced,
   author={Fan, Deng-Ping and Gong, Cheng and Cao, Yang and Ren, Bo and Cheng, Ming-Ming and Borji, Ali},
   title={{Enhanced-alignment Measure for Binary Foreground Map Evaluation}},
   booktitle={IJCAI},
   pages={698--704},
   year={2018}
}
%S-measure
@inproceedings{fan2017structure,
  author    = {Fan, Deng-Ping and Cheng, Ming-Ming and Liu, Yun and Li, Tao and Borji, Ali},
  title     = {{Structure-measure: A New Way to Evaluate Foreground Maps}},
  booktitle = {IEEE ICCV},
  year      = {2017},
  pages     = {4548-4557}
}

##License

Copyright (c) 2019, Deng-Ping Fan, Wenguan Wang, Ming-Ming Cheng, Jianbing Shen.
All rights reserved.

This code is for academic communication only and not for commercial purposes. 
If you want to use for commercial please contact me.

Redistribution and use in source with or without
modification, are permitted provided that the following conditions are
met:
		* Redistributions of source code must retain the above copyright
  		  notice, this list of conditions and the following disclaimer.
		* Redistributions in binary form must reproduce the above copyright
  		  notice, this list of conditions and the following disclaimer in
  		  the documentation and/or other materials provided with the distribution

THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE 	
LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
POSSIBILITY OF SUCH DAMAGE.

If you find any bugs, please contact Deng-Ping Fan ([email protected]).

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].