PGNet: Progressive Feature Guide Learning Network for Three-dimensional Shape Recognition

Author(s):  
Jie Nie ◽  
Zhi-Qiang Wei ◽  
Weizhi Nie ◽  
An-An Liu

Three-dimensional (3D) shape recognition is a popular topic and has potential application value in the field of computer vision. With the recent proliferation of deep learning, various deep learning models have achieved state-of-the-art performance. Among them, multiview-based 3D shape representation has received increased attention in recent years, and related approaches have shown significant improvement in 3D shape recognition. However, these methods focus on feature learning based on the design of the network and ignore the correlation among views. In this article, we propose a novel progressive feature guide learning network (PGNet) that focuses on the correlation among multiple views and integrates multiple modalities for 3D shape recognition. In particular, we propose two information fusion schemes from visual and feature aspects. The visual fusion scheme focuses on the view level and employs the soft-attention model to define the weights of views for visual information fusion. The feature fusion scheme focuses on the feature dimension information and employs the quantified feature as the mask to further optimize the feature. These two schemes jointly construct a PGNet for 3D shape representation. The classic ModelNet40 and ShapeNetCore55 datasets are applied to demonstrate the performance of our approach. The corresponding experiment also demonstrates the superiority of our approach.

Electronics ◽  
2020 ◽  
Vol 9 (4) ◽  
pp. 649
Author(s):  
Long Hoang ◽  
Suk-Hwan Lee ◽  
Ki-Ryong Kwon

3D shape recognition becomes necessary due to the popularity of 3D data resources. This paper aims to introduce the new method, hybrid deep learning network convolution neural network–support vector machine (CNN–SVM), for 3D recognition. The vertices of the 3D mesh are interpolated to be converted into Point Clouds; those Point Clouds are rotated for 3D data augmentation. We obtain and store the 2D projection of this 3D augmentation data in a 32 × 32 × 12 matrix, the input data of CNN–SVM. An eight-layer CNN is used as the algorithm for feature extraction, then SVM is applied for classifying feature extraction. Two big datasets, ModelNet40 and ModelNet10, of the 3D model are used for model validation. Based on our numerical experimental results, CNN–SVM is more accurate and efficient than other methods. The proposed method is 13.48% more accurate than the PointNet method in ModelNet10 and 8.5% more precise than 3D ShapeNets for ModelNet40. The proposed method works with both the 3D model in the augmented/virtual reality system and in the 3D Point Clouds, an output of the LIDAR sensor in autonomously driving cars.


Author(s):  
Haoxuan You ◽  
Yifan Feng ◽  
Xibin Zhao ◽  
Changqing Zou ◽  
Rongrong Ji ◽  
...  

Three-dimensional (3D) shape recognition has drawn much research attention in the field of computer vision. The advances of deep learning encourage various deep models for 3D feature representation. For point cloud and multi-view data, two popular 3D data modalities, different models are proposed with remarkable performance. However the relation between point cloud and views has been rarely investigated. In this paper, we introduce Point-View Relation Network (PVRNet), an effective network designed to well fuse the view features and the point cloud feature with a proposed relation score module. More specifically, based on the relation score module, the point-single-view fusion feature is first extracted by fusing the point cloud feature and each single view feature with point-singe-view relation, then the pointmulti- view fusion feature is extracted by fusing the point cloud feature and the features of different number of views with point-multi-view relation. Finally, the point-single-view fusion feature and point-multi-view fusion feature are further combined together to achieve a unified representation for a 3D shape. Our proposed PVRNet has been evaluated on ModelNet40 dataset for 3D shape classification and retrieval. Experimental results indicate our model can achieve significant performance improvement compared with the state-of-the-art models.


2013 ◽  
Vol 18 (1) ◽  
pp. 23-29
Author(s):  
Dariusz Frejlichowski

Abstract In this paper an algorithm for the representation of 3D models is described and experimentally evaluated. Three-dimensional objects are becoming very popular recently and they are processed in various ways - analysed, retrieved, recognised, and so on. Moreover, they are employed in various aplications, such as virtual reality, entertainment, Internet, Computer Aided Design, or even in biometrics or medical imaging. That is why the development of appropriate algorithms for the representation of 3D objects is so important recently. These algorithms - so called 3D shape descriptors - are assumed to be invariant to particular transformations and deformations. One of the possible approaches is based on the projections of a 3D object into planar shapes and representation of them using a 2D shape descriptor. An algorithm realising this idea is described in this paper. Its first stage is based on the rendering of 20 2D projections, from various points of view. Later, the obtained projections are stored in a form of bitmaps and the Curvature Scale Space algorithm is applied for the description of the planar shapes extracted from them. The proposed approach is experimentally compared with several other 3D shape representation methods.


2017 ◽  
Vol 259 ◽  
pp. 183-193 ◽  
Author(s):  
Shuhui Bu ◽  
Lei Wang ◽  
Pengcheng Han ◽  
Zhenbao Liu ◽  
Ke Li

2014 ◽  
Vol 21 (4) ◽  
pp. 38-46 ◽  
Author(s):  
Shuhui Bu ◽  
Shaoguang Cheng ◽  
Zhenbao Liu ◽  
Junwei Han

IEEE Access ◽  
2021 ◽  
Vol 9 ◽  
pp. 3451-3460
Author(s):  
Shanlin Sun ◽  
Yun Li ◽  
Minjie Ren ◽  
Guo Li ◽  
Xing Yao

Author(s):  
Yong Xu ◽  
Chaoda Zheng ◽  
Ruotao Xua ◽  
Yuhui Quan ◽  
Haibin Ling

Sign in / Sign up

Export Citation Format

Share Document