Thanks to visit codestin.com
Credit goes to github.com

Skip to content

Visual place recognition from opposing viewpoints under extreme appearance variations

License

Notifications You must be signed in to change notification settings

oravus/seq2single

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

4 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Look No Deeper: Recognizing Places from Opposing Viewpoints under Varying Scene Appearance using Single-View Depth Estimation

This is the source code for the paper titled: "Look No Deeper: Recognizing Places from Opposing Viewpoints under Varying Scene Appearance using Single-View Depth Estimation", [arXiv][IEEE Xplore].

If you find this work useful, please cite it as: Garg, S., Babu V, M., Dharmasiri, T., Hausler, S., Suenderhauf, N., Kumar, S., Drummond, T., & Milford, M. (2019). Look no deeper: Recognizing places from opposing viewpoints under varying scene appearance using single-view depth estimation. In IEEE International Conference on Robotics and Automation (ICRA), 2019. IEEE.

bibtex:

@inproceedings{garg2019look,
title={Look No Deeper: Recognizing Places from Opposing Viewpoints under Varying Scene Appearance using Single-View Depth Estimation},
author={Garg, Sourav and Babu V, Madhu and Dharmasiri, Thanuja and Hausler, Stephen and Suenderhauf, Niko and Kumar, Swagat and Drummond, Tom and Milford, Michael},
booktitle={IEEE International Conference on Robotics and Automation (ICRA)},
year={2019}
}

Illustration of the proposed approach

An image depicting topometric representation.

Requirements

  • Ubuntu (Tested on 16.04)
  • Jupyter (Tested on 4.4.0)
  • Python (Tested on 3.5.6)
    • numpy (Tested on 1.15.2)
    • scipy (Tested on 1.1.0)

Optionally, for vis_results.ipynb:

  • Matplotlib (Tested on 2.0.2)

Download an example dataset and its pre-computed representations

  1. In seq2single/precomputed/, download pre-computed representations (~10 GB). Please refer to the seq2single/precomputed/readme.md for instructions on how to compute these representations.

  2. [Optional] In seq2single/images/, download images (~1 GB). These images are a subset of two different traverses from the Oxford Robotcar dataset.

(Note: These download links from Mega.nz require you to first create an account (free))

Run

  1. The Jupyter notebook seq2single.ipynb first loads the pre-computed global image descriptors to find top matches. These matches are re-ranked with the proposed method using the pre-computed depth masks and dense conv5 features.

License

The code is released under MIT License.

Related Projects

Delta Descriptors (2020)

CoarseHash (2020)

LoST (2018)