UnDispNet: Unsupervised Learning for Multi-Stage Monocular Depth Prediction

Author(s):  
Vinay Kaushik ◽  
Brejesh Lall
Author(s):  
Vincent Casser ◽  
Soeren Pirk ◽  
Reza Mahjourian ◽  
Anelia Angelova

Learning to predict scene depth from RGB inputs is a challenging task both for indoor and outdoor robot navigation. In this work we address unsupervised learning of scene depth and robot ego-motion where supervision is provided by monocular videos, as cameras are the cheapest, least restrictive and most ubiquitous sensor for robotics. Previous work in unsupervised image-to-depth learning has established strong baselines in the domain. We propose a novel approach which produces higher quality results, is able to model moving objects and is shown to transfer across data domains, e.g. from outdoors to indoor scenes. The main idea is to introduce geometric structure in the learning process, by modeling the scene and the individual objects; camera ego-motion and object motions are learned from monocular videos as input. Furthermore an online refinement method is introduced to adapt learning on the fly to unknown domains. The proposed approach outperforms all state-of-the-art approaches, including those that handle motion e.g. through learned flow. Our results are comparable in quality to the ones which used stereo as supervision and significantly improve depth prediction on scenes and datasets which contain a lot of object motion. The approach is of practical relevance, as it allows transfer across environments, by transferring models trained on data collected for robot navigation in urban scenes to indoor navigation settings. The code associated with this paper can be found at https://sites.google.com/view/struct2depth.


IEEE Access ◽  
2019 ◽  
Vol 7 ◽  
pp. 77839-77848 ◽  
Author(s):  
Junning Zhang ◽  
Qunxing Su ◽  
Pengyuan Liu ◽  
Chao Xu ◽  
Yanlong Chen

Author(s):  
Zhaokai Wang ◽  
Limin Xiao ◽  
Rongbin Xu ◽  
Shubin Su ◽  
Shupan Li ◽  
...  

2019 ◽  
Vol 11 (3) ◽  
pp. 615-627 ◽  
Author(s):  
Junning Zhang ◽  
Qunxing Su ◽  
Pengyuan Liu ◽  
Chao Xu ◽  
Yanlong Chen

2020 ◽  
Vol 34 (07) ◽  
pp. 12257-12264 ◽  
Author(s):  
Xinlong Wang ◽  
Wei Yin ◽  
Tao Kong ◽  
Yuning Jiang ◽  
Lei Li ◽  
...  

Monocular depth estimation enables 3D perception from a single 2D image, thus attracting much research attention for years. Almost all methods treat foreground and background regions (“things and stuff”) in an image equally. However, not all pixels are equal. Depth of foreground objects plays a crucial role in 3D object recognition and localization. To date how to boost the depth prediction accuracy of foreground objects is rarely discussed. In this paper, we first analyze the data distributions and interaction of foreground and background, then propose the foreground-background separated monocular depth estimation (ForeSeE) method, to estimate the foreground and background depth using separate optimization objectives and decoders. Our method significantly improves the depth estimation performance on foreground objects. Applying ForeSeE to 3D object detection, we achieve 7.5 AP gains and set new state-of-the-art results among other monocular methods. Code will be available at: https://github.com/WXinlong/ForeSeE.


Sign in / Sign up

Export Citation Format

Share Document