scholarly journals Automatic object annotation in streamed and remotely explored large 3D reconstructions

Author(s):  
Benjamin Höller ◽  
Annette Mossel ◽  
Hannes Kaufmann

AbstractWe introduce a novel framework for 3D scene reconstruction with simultaneous object annotation, using a pre-trained 2D convolutional neural network (CNN), incremental data streaming, and remote exploration, with a virtual reality setup. It enables versatile integration of any 2D box detection or segmentation network. We integrate new approaches to (i) asynchronously perform dense 3D-reconstruction and object annotation at interactive frame rates, (ii) efficiently optimize CNN results in terms of object prediction and spatial accuracy, and (iii) generate computationally-efficient colliders in large triangulated 3D-reconstructions at run-time for 3D scene interaction. Our method is novel in combining CNNs with long and varying inference time with live 3D-reconstruction from RGB-D camera input. We further propose a lightweight data structure to store the 3D-reconstruction data and object annotations to enable fast incremental data transmission for real-time exploration with a remote client, which has not been presented before. Our framework achieves update rates of 22 fps (SSD Mobile Net) and 19 fps (Mask RCNN) for indoor environments up to 800 m3. We evaluated the accuracy of 3D-object detection. Our work provides a versatile foundation for semantic scene understanding of large streamed 3D-reconstructions, while being independent from the CNN’s processing time. Source code is available for non-commercial use.

2020 ◽  
Vol 34 (07) ◽  
pp. 11402-11409
Author(s):  
Siqi Li ◽  
Changqing Zou ◽  
Yipeng Li ◽  
Xibin Zhao ◽  
Yue Gao

This paper presents an end-to-end 3D convolutional network named attention-based multi-modal fusion network (AMFNet) for the semantic scene completion (SSC) task of inferring the occupancy and semantic labels of a volumetric 3D scene from single-view RGB-D images. Compared with previous methods which use only the semantic features extracted from RGB-D images, the proposed AMFNet learns to perform effective 3D scene completion and semantic segmentation simultaneously via leveraging the experience of inferring 2D semantic segmentation from RGB-D images as well as the reliable depth cues in spatial dimension. It is achieved by employing a multi-modal fusion architecture boosted from 2D semantic segmentation and a 3D semantic completion network empowered by residual attention blocks. We validate our method on both the synthetic SUNCG-RGBD dataset and the real NYUv2 dataset and the results show that our method respectively achieves the gains of 2.5% and 2.6% on the synthetic SUNCG-RGBD dataset and the real NYUv2 dataset against the state-of-the-art method.


2014 ◽  
Vol 998-999 ◽  
pp. 1018-1023
Author(s):  
Rui Bin Guo ◽  
Tao Guan ◽  
Dong Xiang Zhou ◽  
Ke Ju Peng ◽  
Wei Hong Fan

Recent approaches for reconstructing 3D scenes from image collections only produce single scene models. To build a unified scene model that contains multiple subsets, we present a novel method for registration of 3D scene reconstructions in different scales. It first normalizes the scales of the models building on similarity reconstruction by the constraint of the 3D position of shared cameras. Then we use Cayley transform to fit the matrix of coordinates transformation for the models in normalization scales. The experimental results show the effectiveness and scalability of the proposed approach.


2004 ◽  
Vol 43 (04) ◽  
pp. 320-326 ◽  
Author(s):  
T. Schüle ◽  
C. Schnörr ◽  
J. Hornegger ◽  
S. Weber

Summary Objectives: We investigate the feasibility of binary-valued 3D tomographic reconstruction using only a small number of projections acquired over a limited range of angles. Methods: Regularization of this strongly ill-posed problem is achieved by (i) confining the reconstruction to binary vessel/non-vessel decisions, and (ii) by minimizing a global functional involving a smoothness prior. Results: Our approach successfully reconstructs volumetric vessel structures from three projections taken within 90°. The percentage of reconstructed voxels differing from ground truth is below 1%. Conclusion: We demonstrate that for particular applications – like Digital Subtraction Angiography – 3D reconstructions are possible where conventional methods must fail, due to a severely limited imaging geometry. This could play an important role for dose reduction and 3D reconstruction using non-conventional technical setups.


2019 ◽  
Vol 10 (1) ◽  
Author(s):  
Julián Tachella ◽  
Yoann Altmann ◽  
Nicolas Mellado ◽  
Aongus McCarthy ◽  
Rachael Tobin ◽  
...  

Abstract Single-photon lidar has emerged as a prime candidate technology for depth imaging through challenging environments. Until now, a major limitation has been the significant amount of time required for the analysis of the recorded data. Here we show a new computational framework for real-time three-dimensional (3D) scene reconstruction from single-photon data. By combining statistical models with highly scalable computational tools from the computer graphics community, we demonstrate 3D reconstruction of complex outdoor scenes with processing times of the order of 20 ms, where the lidar data was acquired in broad daylight from distances up to 320 metres. The proposed method can handle an unknown number of surfaces in each pixel, allowing for target detection and imaging through cluttered scenes. This enables robust, real-time target reconstruction of complex moving scenes, paving the way for single-photon lidar at video rates for practical 3D imaging applications.


Author(s):  
R. Hegerl ◽  
Z. Cejka ◽  
W. Baumeister

Two major questions concerning the results of 3D reconstruction are: 1) The effects of the interaction between specimen and specimen support and 2) the effects of stain level and stain distribution. These questions were approached using the S-layer of the bacterium Clostridium thermohydrosulphuricum which in the electron microscope is usually doubled upon itself. From these specimens independent 3D reconstructions of the upper and lower layers were made and compared with each other.


Sign in / Sign up

Export Citation Format

Share Document