All Projects → jiwei0921 → Dmra

jiwei0921 / Dmra

Licence: mit
Code and Dataset for ICCV 2019 paper. "Depth-induced Multi-scale Recurrent Attention Network for Saliency Detection".

Programming Languages

python
139335 projects - #7 most used programming language

Labels

Projects that are alternatives of or similar to Dmra

DPANet
DPANet : Depth Potentiality-Aware Gated Attention Network for RGB-D Salient Object Detection
Stars: ✭ 22 (-71.05%)
Mutual labels:  rgbd
Intrinsic3d
Intrinsic3D - High-Quality 3D Reconstruction by Joint Appearance and Geometry Optimization with Spatially-Varying Lighting (ICCV 2017)
Stars: ✭ 297 (+290.79%)
Mutual labels:  rgbd
3dmatch Toolbox
3DMatch - a 3D ConvNet-based local geometric descriptor for aligning 3D meshes and point clouds.
Stars: ✭ 571 (+651.32%)
Mutual labels:  rgbd
DeT
Dataset and Code for the paper "DepthTrack: Unveiling the Power of RGBD Tracking" (ICCV2021), and "Depth-only Object Tracking" (BMVC2021)
Stars: ✭ 39 (-48.68%)
Mutual labels:  rgbd
Apc Vision Toolbox
MIT-Princeton Vision Toolbox for the Amazon Picking Challenge 2016 - RGB-D ConvNet-based object segmentation and 6D object pose estimation.
Stars: ✭ 277 (+264.47%)
Mutual labels:  rgbd
Co Fusion
Co-Fusion: Real-time Segmentation, Tracking and Fusion of Multiple Objects
Stars: ✭ 400 (+426.32%)
Mutual labels:  rgbd
GeobitNonrigidDescriptor ICCV 2019
C++ implementation of the nonrigid descriptor Geobit presented at ICCV 2019 "GEOBIT: A Geodesic-Based Binary Descriptor Invariant to Non-Rigid Deformations for RGB-D Images"
Stars: ✭ 11 (-85.53%)
Mutual labels:  rgbd
Openni2 camera
ROS wrapper for openni 2.0
Stars: ✭ 36 (-52.63%)
Mutual labels:  rgbd
Open3d Ml
An extension of Open3D to address 3D Machine Learning tasks
Stars: ✭ 284 (+273.68%)
Mutual labels:  rgbd
Tsdf Fusion Python
Python code to fuse multiple RGB-D images into a TSDF voxel volume.
Stars: ✭ 464 (+510.53%)
Mutual labels:  rgbd
staticfusion
StaticFusion
Stars: ✭ 107 (+40.79%)
Mutual labels:  rgbd
TorchSSC
Implement some state-of-the-art methods of Semantic Scene Completion (SSC) task in PyTorch. [1] 3D Sketch-aware Semantic Scene Completion via Semi-supervised Structure Prior (CVPR 2020)
Stars: ✭ 35 (-53.95%)
Mutual labels:  rgbd
Maskfusion
MaskFusion: Real-Time Recognition, Tracking and Reconstruction of Multiple Moving Objects
Stars: ✭ 404 (+431.58%)
Mutual labels:  rgbd
maplab realsense
Simple ROS wrapper for the Intel RealSense driver with a focus on the ZR300.
Stars: ✭ 22 (-71.05%)
Mutual labels:  rgbd
Cilantro
A lean C++ library for working with point cloud data
Stars: ✭ 577 (+659.21%)
Mutual labels:  rgbd
referit3d
Code accompanying our ECCV-2020 paper on 3D Neural Listeners.
Stars: ✭ 59 (-22.37%)
Mutual labels:  rgbd
Handeye calib camodocal
Easy to use and accurate hand eye calibration which has been working reliably for years (2016-present) with kinect, kinectv2, rgbd cameras, optical trackers, and several robots including the ur5 and kuka iiwa.
Stars: ✭ 364 (+378.95%)
Mutual labels:  rgbd
Peac
Fast Plane Extraction Using Agglomerative Hierarchical Clustering (AHC)
Stars: ✭ 51 (-32.89%)
Mutual labels:  rgbd
Scannet
Stars: ✭ 860 (+1031.58%)
Mutual labels:  rgbd
Tsdf Fusion
Fuse multiple depth frames into a TSDF voxel volume.
Stars: ✭ 426 (+460.53%)
Mutual labels:  rgbd

DMRA_RGBD-SOD

Code repository for our paper entilted "Depth-induced Multi-scale Recurrent Attention Network for Saliency Detection" accepted at ICCV 2019 (poster).

Overall

avatar

The proposed Dataset

  • Dataset: DUTLF
  1. This dataset consists of DUTLF-MV, DUTLF-FS, DUTLF-Depth.
  2. The dataset will be expanded to 3000 about real scenes.
  3. We are working on it and will make it publicly available soon.
  • Dataset: DUTLF-Depth
  1. The dataset is part of DUTLF dataset captured by Lytro camera, and we selected a more accurate 1200 depth map pairs for more accurate RGB-D saliency detection.
  2. We create a large scale RGB-D dataset(DUTLF-Depth) with 1200 paired images containing more complex scenarios, such as multiple or transparent objects, similar foreground and background, complex background, low-intensity environment. This challenging dataset can contribute to comprehensively evaluating saliency models.

avatar

  • The dataset link can be found here. And we split the dataset including 800 training set and 400 test set.

DMRA Code

> Requirment

  • pytorch 0.3.0+
  • torchvision
  • PIL
  • numpy

> Usage

1. Clone the repo

git clone https://github.com/jiwei0921/DMRA.git
cd DMRA/

2. Train/Test

  • test
    Download related dataset link, and set the param '--phase' as "test" and '--param' as 'True' in demo.py. Meanwhile, you need to set dataset path and checkpoint name correctly.
python demo.py
  • train
    Our train-augment dataset link [ fetch code haxl ] / train-ori dataset, and set the param '--phase' as "train" and '--param' as 'True'(loading checkpoint) or 'False'(no loading checkpoint) in demo.py. Meanwhile, you need to set dataset path and checkpoint name correctly.
python demo.py

> Training info and pre-trained models for DMRA

To better understand, we retrain our network and record some detailed training details as well as corresponding pre-trained models.

Iterations Loss NJUD(F-measure) NJUD(MAE) NLPR(F-measure) NLPR(MAE) download link
100W 958 0.882 0.048 0.867 0.031 link
70W 2413 0.876 0.050 0.854 0.033 link
40W 3194 0.861 0.056 0.823 0.037 link
16W 8260 0.805 0.081 0.725 0.056 link
2W 33494 0.009 0.470 0.030 0.452 link
0W 45394 - - - - -
  • Tips: The results of the paper shall prevail. Because of the randomness of the training process, the results fluctuated slightly.

> Results

| DUTLF-Depth | | NJUD | | NLPR | | STEREO | | LFSD | | RGBD135 | | SSD |

  • Note: For evaluation, all results are implemented on this ready-to-use toolbox.
  • SIP results: This is test results on SIP dataset, and fetch code is 'fi5h'.

> Related RGB-D Saliency Datasets

All common RGB-D Saliency Datasets we collected are shared in ready-to-use manner.

  • The web link is here.

If you think this work is helpful, please cite

@inproceedings{piao2019depth,
  title={Depth-induced multi-scale recurrent attention network for saliency detection},
  author={Piao, Yongri and Ji, Wei and Li, Jingjing and Zhang, Miao and Lu, Huchuan},
  booktitle={Proceedings of the IEEE International Conference on Computer Vision},
  pages={7254--7263},
  year={2019}
}

Related SOTA RGB-D methods' results on our dataset

Meanwhile, we also provide other state-of-the-art RGB-D methods' results on our proposed dataset, and you can directly download their results (All results,2gs2).

No. Pub. Name Title Download
14 ICCV2019 DMRA Depth-induced multi-scale recurrent attention network for saliency detection results, g7rz
13 CVPR2019 CPFP Depth-induced multi-scale recurrent attention network for saliency detection results, g7rz
12 TIP2019 TANet Three-stream attention-aware network for RGB-D salient object detection results, g7rz
11 PR2019 MMCI Multi-modal fusion network with multiscale multi-path and cross-modal interactions for RGB-D salient object detection results, g7rz
10 ICME2019 PDNet Pdnet: Prior-model guided depth-enhanced network for salient object detection results, g7rz
09 CVPR2018 PCA Progressively Complementarity-Aware Fusion Network for RGB-D Salient Object Detection results, g7rz
08 ICCVW2017 CDCP An innovative salient object detection using center-dark channel prior results, g7rz
07 TCyb2017 CTMF CNNs-based RGB-D saliency detection via cross-view transfer and multiview fusion results, g7rz
06 TIP2017 DF RGBD salient object detection via deep fusion results, g7rz
05 CAIP2017 MB A Multilayer Backpropagation Saliency Detection Algorithm Based on Depth Mining results, g7rz
04 SPL2016 DCMC Saliency detection for stereoscopic images based on depth confidence analysis and multiple cues fusion results, g7rz
03 ECCV2014 LHM-NLPR Rgbd salient object detection: a benchmark and algorithms results, g7rz
02 ICIP2014 ACSD Depth saliency based on anisotropic center-surround difference results, g7rz
01 ICIMCS2014 DES Depth enhanced saliency detection method results, g7rz

Contact Us

If you have any questions, please contact us ( [email protected] or [email protected] ).

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].