An Improved Multi-View Convolutional Neural Network for 3D Object Retrieval

2020 ◽  
Vol 29 ◽  
pp. 7917-7930
Author(s):  
Xinwei He ◽  
Song Bai ◽  
Jiajia Chu ◽  
Xiang Bai
Author(s):  
Zhengyue Huang ◽  
Zhehui Zhao ◽  
Hengguang Zhou ◽  
Xibin Zhao ◽  
Yue Gao

3D object retrieval has a compelling demand in the field of computer vision with the rapid development of 3D vision technology and increasing applications of 3D objects. 3D objects can be described in different ways such as voxel, point cloud, and multi-view. Among them, multi-view based approaches proposed in recent years show promising results. Most of them require a fixed predefined camera position setting which provides a complete and uniform sampling of views for objects in the training stage. However, this causes heavy over-fitting problems which make the models failed to generalize well in free camera setting applications, particularly when insufficient views are provided. Experiments show the performance drastically drops when the number of views reduces, hindering these methods from practical applications. In this paper, we investigate the over-fitting issue and remove the constraint of the camera setting. First, two basic feature augmentation strategies Dropout and Dropview are introduced to solve the over-fitting issue, and a more precise and more efficient method named DropMax is proposed after analyzing the drawback of the basic ones. Then, by reducing the over-fitting issue, a camera constraint-free multi-view convolutional neural network named DeepCCFV is constructed. Extensive experiments on both single-modal and cross-modal cases demonstrate the effectiveness of the proposed method in free camera settings comparing with existing state-of-theart 3D object retrieval methods.


2005 ◽  
Vol 41 (4) ◽  
pp. 179 ◽  
Author(s):  
J.-L. Shih ◽  
C.-H. Lee ◽  
J.T. Wang

2014 ◽  
Vol 21 (3) ◽  
pp. 52-57 ◽  
Author(s):  
Yue Gao ◽  
Qionghai Dai

2015 ◽  
Vol 76 (3) ◽  
pp. 4091-4104 ◽  
Author(s):  
Weizhi Nie ◽  
Xixi Li ◽  
Anan Liu ◽  
Yuting Su

Author(s):  
Zhiyong Gao ◽  
Jianhong Xiang

Background: While detecting the object directly from the 3D point cloud, the natural 3D patterns and invariance of 3D data are often obscure. Objective: In this work, we aimed at studying the 3D object detection from discrete, disordered and sparse 3D point clouds. Methods: The CNN is composed of the frustum sequence module, 3D instance segmentation module S-NET, 3D point cloud transformation module T-NET, and 3D boundary box estimation module E-NET. The search space of the object is determined by the frustum sequence module. The instance segmentation of the point cloud is performed by the 3D instance segmentation module. The 3D coordinates of the object are confirmed by the transformation module and the 3D bounding box estimation module. Results: Evaluated on KITTI benchmark dataset, our method outperforms the state of the art by remarkable margins while having real-time capability. Conclusion: We achieve real-time 3D object detection by proposing an improved convolutional neural network (CNN) based on image-driven point clouds.


Author(s):  
Ilyass Ouazzani Taybi ◽  
Rachid Alaoui ◽  
Fatima Rafii Zakani ◽  
Khadija Arhid ◽  
Mohcine Bouksim ◽  
...  

Sign in / Sign up

Export Citation Format

Share Document