Share Email Print

Proceedings Paper

Fast scene layout estimation via deep hashing
Author(s): Yi Zhu; Wenbing Luo; Hanxi Li; Mingwen Wang
Format Member Price Non-Member Price
PDF $14.40 $18.00
cover GOOD NEWS! Your organization subscribes to the SPIE Digital Library. You may be able to download this paper for free. Check Access

Paper Abstract

In this work, we propose an efficient method for accurately estimating the scene layout in both outdoor and indoor scenarios. For outdoor scenes, the horizon line in a road image is estimated while for indoor scenes, the wall-wall, wallceiling and wall-floor edges are estimated. A number of image patches are first cropped from the image and then feed into a convolution neural network which is originally trained for object detection. The yielded deep features from three different layers are compared with the features of the training patches, in a spatial-aware hashing fashion. The horizon line is then estimated via a sophisticated voting stage in which different voters are considered differently according to their importances. In particular, for the more complex labels (in indoor scenes), we introduce the structural forest for further enhancing the deep features before learning the hashing function. In practice, the proposed algorithm outperforms the state-of-the-art methods in accuracy for outdoor scenes while achieves the comparable performance to the best indoor scene layout estimators. Further more, the proposed method is real-time speed (up to 25 fps).

Paper Details

Date Published: 26 July 2018
PDF: 10 pages
Proc. SPIE 10828, Third International Workshop on Pattern Recognition, 1082814 (26 July 2018); doi: 10.1117/12.2501793
Show Author Affiliations
Yi Zhu, Jiangxi Normal Univ. (China)
Wenbing Luo, Jiangxi Normal Univ. (China)
Hanxi Li, Jiangxi Normal Univ. (China)
Mingwen Wang, Jiangxi Normal Univ. (China)

Published in SPIE Proceedings Vol. 10828:
Third International Workshop on Pattern Recognition
Xudong Jiang; Zhenxiang Chen; Guojian Chen, Editor(s)

© SPIE. Terms of Use
Back to Top