Scene Reconstruction from a Single Depth Image Using 3D CNN

Author(s):  
Alessandro Palla ◽  
David Moloney ◽  
Luca Fanucci
2017 ◽  
Vol 39 (6) ◽  
pp. 106-121
Author(s):  
A. O. Verpahovskaya ◽  
V. N. Pilipenko ◽  
Е. V. Pylypenko

2020 ◽  
Vol 25 (3) ◽  
pp. 265-276
Author(s):  
K.M. Shepilova ◽  
◽  
A.V. Sotnikov ◽  
A.V. Shipatov ◽  
Yu.V. Savchenko ◽  
...  

2012 ◽  
Vol 38 (9) ◽  
pp. 1428 ◽  
Author(s):  
Xin LIU ◽  
Feng-Mei SUN ◽  
Zhan-Yi HU

2019 ◽  
Author(s):  
Han-Chi Hsieh ◽  
Wei-Zhong Zheng ◽  
Ko-Chiang Chen ◽  
Ying-Hui Lai

2021 ◽  
Vol 40 (3) ◽  
pp. 1-12
Author(s):  
Hao Zhang ◽  
Yuxiao Zhou ◽  
Yifei Tian ◽  
Jun-Hai Yong ◽  
Feng Xu

Reconstructing hand-object interactions is a challenging task due to strong occlusions and complex motions. This article proposes a real-time system that uses a single depth stream to simultaneously reconstruct hand poses, object shape, and rigid/non-rigid motions. To achieve this, we first train a joint learning network to segment the hand and object in a depth image, and to predict the 3D keypoints of the hand. With most layers shared by the two tasks, computation cost is saved for the real-time performance. A hybrid dataset is constructed here to train the network with real data (to learn real-world distributions) and synthetic data (to cover variations of objects, motions, and viewpoints). Next, the depth of the two targets and the keypoints are used in a uniform optimization to reconstruct the interacting motions. Benefitting from a novel tangential contact constraint, the system not only solves the remaining ambiguities but also keeps the real-time performance. Experiments show that our system handles different hand and object shapes, various interactive motions, and moving cameras.


Author(s):  
Alexander Bigalke ◽  
Lasse Hansen ◽  
Jasper Diesel ◽  
Mattias P. Heinrich

Abstract Purpose Body weight is a crucial parameter for patient-specific treatments, particularly in the context of proper drug dosage. Contactless weight estimation from visual sensor data constitutes a promising approach to overcome challenges arising in emergency situations. Machine learning-based methods have recently been shown to perform accurate weight estimation from point cloud data. The proposed methods, however, are designed for controlled conditions in terms of visibility and position of the patient, which limits their practical applicability. In this work, we aim to decouple accurate weight estimation from such specific conditions by predicting the weight of covered patients from voxelized point cloud data. Methods We propose a novel deep learning framework, which comprises two 3D CNN modules solving the given task in two separate steps. First, we train a 3D U-Net to virtually uncover the patient, i.e. to predict the patient’s volumetric surface without a cover. Second, the patient’s weight is predicted from this 3D volume by means of a 3D CNN architecture, which we optimized for weight regression. Results We evaluate our approach on a lying pose dataset (SLP) under two different cover conditions. The proposed framework considerably improves on the baseline model by up to $${16}{\%}$$ 16 % and reduces the gap between the accuracy of weight estimates for covered and uncovered patients by up to $${52}{\%}$$ 52 % . Conclusion We present a novel pipeline to estimate the weight of patients, which are covered by a blanket. Our approach relaxes the specific conditions that were required for accurate weight estimates by previous contactless methods and thus constitutes an important step towards fully automatic weight estimation in clinical practice.


Sign in / Sign up

Export Citation Format

Share Document