All Projects → oravus → seq2single

oravus / seq2single

Licence: MIT License
Visual place recognition from opposing viewpoints under extreme appearance variations

Programming Languages

Jupyter Notebook
11667 projects

Projects that are alternatives of or similar to seq2single

lostX
(RSS 2018) LoST - Visual Place Recognition using Visual Semantics for Opposite Viewpoints across Day and Night
Stars: ✭ 60 (+300%)
Mutual labels:  place-recognition, day-night-cycle, visual-place-recognition
Depth estimation
Deep learning model to estimate the depth of image.
Stars: ✭ 62 (+313.33%)
Mutual labels:  depth-estimation
pais-mvs
Multi-view stereo image-based 3D reconstruction
Stars: ✭ 55 (+266.67%)
Mutual labels:  depth-estimation
understanding apr
No description or website provided.
Stars: ✭ 38 (+153.33%)
Mutual labels:  visual-localization
project-defude
Refocus an image just by clicking on it with no additional data
Stars: ✭ 69 (+360%)
Mutual labels:  depth-estimation
Structured-Light-Depth-Acquisition
Matlab Implementation of a 3D Reconstruction algorithm
Stars: ✭ 48 (+220%)
Mutual labels:  depth-estimation
DSGN
DSGN: Deep Stereo Geometry Network for 3D Object Detection (CVPR 2020)
Stars: ✭ 276 (+1740%)
Mutual labels:  depth-estimation
DE resnet unet hyb
Depth estimation from RGB images using fully convolutional neural networks.
Stars: ✭ 40 (+166.67%)
Mutual labels:  depth-estimation
Indoor-SfMLearner
[ECCV'20] Patch-match and Plane-regularization for Unsupervised Indoor Depth Estimation
Stars: ✭ 115 (+666.67%)
Mutual labels:  depth-estimation
improving segmentation with selfsupervised depth
[CVPR21] Implementation of our work "Three Ways to Improve Semantic Segmentation with Self-Supervised Depth Estimation"
Stars: ✭ 189 (+1160%)
Mutual labels:  depth-estimation
sc depth pl
Pytorch Lightning Implementation of SC-Depth (V1, V2...) for Unsupervised Monocular Depth Estimation.
Stars: ✭ 86 (+473.33%)
Mutual labels:  depth-estimation
Semantic-Mono-Depth
Geometry meets semantics for semi-supervised monocular depth estimation - ACCV 2018
Stars: ✭ 98 (+553.33%)
Mutual labels:  depth-estimation
motion trace bulk
MMDモーショントレース自動化一括処理バッチ
Stars: ✭ 36 (+140%)
Mutual labels:  depth-estimation
DiverseDepth
The code and data of DiverseDepth
Stars: ✭ 150 (+900%)
Mutual labels:  depth-estimation
HoHoNet
"HoHoNet: 360 Indoor Holistic Understanding with Latent Horizontal Features" official pytorch implementation.
Stars: ✭ 65 (+333.33%)
Mutual labels:  depth-estimation
FAST LIO SLAM
LiDAR SLAM = FAST-LIO + Scan Context
Stars: ✭ 183 (+1120%)
Mutual labels:  place-recognition
OMG Depth Fusion
Probabilistic depth fusion based on Optimal Mixture of Gaussians for depth cameras
Stars: ✭ 52 (+246.67%)
Mutual labels:  depth-estimation
SupervisedDepthPrediction
Pytorch framework for supervised depth prediction
Stars: ✭ 36 (+140%)
Mutual labels:  depth-estimation
rectified-features
[ECCV 2020] Single image depth prediction allows us to rectify planar surfaces in images and extract view-invariant local features for better feature matching
Stars: ✭ 57 (+280%)
Mutual labels:  depth-estimation
ONNX-HITNET-Stereo-Depth-estimation
Python scripts form performing stereo depth estimation using the HITNET model in ONNX.
Stars: ✭ 21 (+40%)
Mutual labels:  depth-estimation

Look No Deeper: Recognizing Places from Opposing Viewpoints under Varying Scene Appearance using Single-View Depth Estimation

This is the source code for the paper titled: "Look No Deeper: Recognizing Places from Opposing Viewpoints under Varying Scene Appearance using Single-View Depth Estimation", [arXiv][IEEE Xplore].

If you find this work useful, please cite it as: Garg, S., Babu V, M., Dharmasiri, T., Hausler, S., Suenderhauf, N., Kumar, S., Drummond, T., & Milford, M. (2019). Look no deeper: Recognizing places from opposing viewpoints under varying scene appearance using single-view depth estimation. In IEEE International Conference on Robotics and Automation (ICRA), 2019. IEEE.

bibtex:

@inproceedings{garg2019look,
title={Look No Deeper: Recognizing Places from Opposing Viewpoints under Varying Scene Appearance using Single-View Depth Estimation},
author={Garg, Sourav and Babu V, Madhu and Dharmasiri, Thanuja and Hausler, Stephen and Suenderhauf, Niko and Kumar, Swagat and Drummond, Tom and Milford, Michael},
booktitle={IEEE International Conference on Robotics and Automation (ICRA)},
year={2019}
}

Illustration of the proposed approach

An image depicting topometric representation.

Requirements

  • Ubuntu (Tested on 16.04)
  • Jupyter (Tested on 4.4.0)
  • Python (Tested on 3.5.6)
    • numpy (Tested on 1.15.2)
    • scipy (Tested on 1.1.0)

Optionally, for vis_results.ipynb:

  • Matplotlib (Tested on 2.0.2)

Download an example dataset and its pre-computed representations

  1. In seq2single/precomputed/, download pre-computed representations (~10 GB). Please refer to the seq2single/precomputed/readme.md for instructions on how to compute these representations.

  2. [Optional] In seq2single/images/, download images (~1 GB). These images are a subset of two different traverses from the Oxford Robotcar dataset.

(Note: These download links from Mega.nz require you to first create an account (free))

Run

  1. The Jupyter notebook seq2single.ipynb first loads the pre-computed global image descriptors to find top matches. These matches are re-ranked with the proposed method using the pre-computed depth masks and dense conv5 features.

License

The code is released under MIT License.

Related Projects

Delta Descriptors (2020)

CoarseHash (2020)

LoST (2018)

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].