wang-chen / Interestingness
Programming Languages
Labels
Projects that are alternatives of or similar to Interestingness
Visual Interestingness
-
Refer to the project description for more details.
-
This code is for the following paper, which is selected for oral presentation (2%) at ECCV 2020.
Chen Wang, Wenshan Wang, Yuheng Qiu, Yafei Hu, and Sebastian Scherer, Visual Memorability for Robotic Interestingness via Unsupervised Online Learning, European Conference on Computer Vision (ECCV), 2020.
-
We also provide ROS wrapper for this project, you may go to interestingness_ros.
-
You can find the slides on OneDrive.
-
You can find the SubT dataset and the evaluation tools.
Install Dependencies
This version is tested in PyTorch 1.6 (1.7 should also be fine)
pip3 install -r requirements.txt
Long-term Learning
-
You may skip this step, if you download the pre-trained at.pt into folder "saves".
-
Download coco dataset into folder [data-root]:
bash download_coco.sh [data-root] # replace [data-root] by your desired location
The dataset will be look like:
data-root ├──coco ├── annotations │ ├── annotations_trainval2017 │ └── image_info_test2017 └── images ├── test2017 ├── train2017 └── val2017
-
Run
python3 train_coder.py --data-root [data-root] --model-save saves/ae.pt # This requires a long time for training on single GPU. # Create a folder "saves" manually and a model named "ae.pt" will be saved.
Short-term Learning
-
Dowload the SubT front camera data (SubTF) and put into folder "data-root", so that it looks like:
data-root ├──SubTF ├── 0817-ugv0-tunnel0 ├── 0817-ugv1-tunnel0 ├── 0818-ugv0-tunnel1 ├── 0818-ugv1-tunnel1 ├── 0820-ugv0-tunnel1 ├── 0821-ugv0-tunnel0 ├── 0821-ugv1-tunnel0 ├── ground-truth └── train
-
Run
python3 train_interest.py --data-root [data-root] --model-save saves/ae.pt --dataset SubTF --memory-size 1000 --save-flag n1000 # This will read the previous model "ae.pt". # A new model "ae.pt.SubTF.n1000.mse" will be generated.
-
You may skip this step, if you download the pre-trained ae.pt.SubTF.n1000.mse into folder "saves".
On-line Learning
-
Run
python3 test_interest.py --data-root [data-root] --model-save saves/ae.pt.SubTF.n1000.mse --dataset SubTF --test-data 0 # --test-data The sequence ID in the dataset SubTF, [0-6] is avaiable # This will read the trained model "ae.pt.SubTF.n1000.mse" from short-term learning.
-
Alternatively, you may test all sequences by running
bash test.sh
-
This will generate results files in folder "results".
-
You may skip this step, if you download our generated results.
Evaluation
-
We follow the SubT tutorial for evaluation, simply run
python performance.py --data-root [data-root] --save-flag n1000 --category interest-1 # mean accuracy: [0.66235087 0.84281507 0.95655934] python performance.py --data-root [data-root] --save-flag n1000 --category interest-2 # mean accuracy: [0.40703316 0.58456123 0.76820896]
-
This will generate performance figures and create data curves for two categories in folder "performance".
Citation
@inproceedings{wang2020visual,
title={Visual memorability for robotic interestingness via unsupervised online learning},
author={Wang, Chen and Wang, Wenshan and Qiu, Yuheng and Hu, Yafei and Scherer, Sebastian},
booktitle={European Conference on Computer Vision (ECCV)},
year={2020},
organization={Springer}
}
- Download this paper.
You may watch the following video to catch the idea of this work.