scholarly journals A Robust and Reliable Point Cloud Recognition Network Under Rigid Transformation

Author(s):  
Dongrui Liu ◽  
Chuanchuan Chen ◽  
Changqing Xu ◽  
Qi Cai ◽  
Lei Chu ◽  
...  
Sensors ◽  
2021 ◽  
Vol 21 (17) ◽  
pp. 5778
Author(s):  
Baifan Chen ◽  
Hong Chen ◽  
Baojun Song ◽  
Grace Gong

Three-dimensional point cloud registration (PCReg) has a wide range of applications in computer vision, 3D reconstruction and medical fields. Although numerous advances have been achieved in the field of point cloud registration in recent years, large-scale rigid transformation is a problem that most algorithms still cannot effectively handle. To solve this problem, we propose a point cloud registration method based on learning and transform-invariant features (TIF-Reg). Our algorithm includes four modules, which are the transform-invariant feature extraction module, deep feature embedding module, corresponding point generation module and decoupled singular value decomposition (SVD) module. In the transform-invariant feature extraction module, we design TIF in SE(3) (which means the 3D rigid transformation space) which contains a triangular feature and local density feature for points. It fully exploits the transformation invariance of point clouds, making the algorithm highly robust to rigid transformation. The deep feature embedding module embeds TIF into a high-dimension space using a deep neural network, further improving the expression ability of features. The corresponding point cloud is generated using an attention mechanism in the corresponding point generation module, and the final transformation for registration is calculated in the decoupled SVD module. In an experiment, we first train and evaluate the TIF-Reg method on the ModelNet40 dataset. The results show that our method keeps the root mean squared error (RMSE) of rotation within 0.5∘ and the RMSE of translation error close to 0 m, even when the rotation is up to [−180∘, 180∘] or the translation is up to [−20 m, 20 m]. We also test the generalization of our method on the TUM3D dataset using the model trained on Modelnet40. The results show that our method’s errors are close to the experimental results on Modelnet40, which verifies the good generalization ability of our method. All experiments prove that the proposed method is superior to state-of-the-art PCReg algorithms in terms of accuracy and complexity.


Author(s):  
R. Huang ◽  
W. Yao ◽  
Z. Ye ◽  
Y. Xu ◽  
U. Stilla

Abstract. Registration of point clouds is a fundamental problem in the community of photogrammetry and 3D computer vision. Generally, point cloud registration consists of two steps: the search of correspondences and the estimation of transformation parameters. However, to find correspondences from point clouds, generating robust and discriminative features is of necessity. In this paper, we address the problem of extracting robust rotation-invariant features for fast coarse registration of point clouds under the assumption that the pairwise point clouds are transformed with rigid transformation. With a Fourier-based descriptor, point clouds represented by volumetric images can be mapped from the image to feature space. It is achieved by considering a gradient histogram as a continuous angular signal which can be well represented by the spherical harmonics. The rotation-invariance is established based on the Fourier-based analysis, in which high-frequency signals can be filtered out. This makes the extracted features robust to noises and outliers. Then, with the extracted features, pairwise correspondence can be found by the fast search. Finally, the transformation parameters can be estimated by fitting the rigid transformation model using the corresponding points and RANSAC algorithm. Experiments are conducted to prove the effectiveness of our proposed method in the task of point cloud registration. Regarding the experimental results of the point cloud registration using two TLS benchmark point cloud datasets, featuring with limited overlaps and uneven point densities and covering different urban scenes, our proposed method can achieve a fast coarse registration with rotation errors of less than 1 degree and translation errors of less than 1m.


2016 ◽  
Vol 136 (8) ◽  
pp. 1078-1084
Author(s):  
Shoichi Takei ◽  
Shuichi Akizuki ◽  
Manabu Hashimoto

Author(s):  
Jiayong Yu ◽  
Longchen Ma ◽  
Maoyi Tian, ◽  
Xiushan Lu

The unmanned aerial vehicle (UAV)-mounted mobile LiDAR system (ULS) is widely used for geomatics owing to its efficient data acquisition and convenient operation. However, due to limited carrying capacity of a UAV, sensors integrated in the ULS should be small and lightweight, which results in decrease in the density of the collected scanning points. This affects registration between image data and point cloud data. To address this issue, the authors propose a method for registering and fusing ULS sequence images and laser point clouds, wherein they convert the problem of registering point cloud data and image data into a problem of matching feature points between the two images. First, a point cloud is selected to produce an intensity image. Subsequently, the corresponding feature points of the intensity image and the optical image are matched, and exterior orientation parameters are solved using a collinear equation based on image position and orientation. Finally, the sequence images are fused with the laser point cloud, based on the Global Navigation Satellite System (GNSS) time index of the optical image, to generate a true color point cloud. The experimental results show the higher registration accuracy and fusion speed of the proposed method, thereby demonstrating its accuracy and effectiveness.


2020 ◽  
Vol 28 (7) ◽  
pp. 1618-1625
Author(s):  
Fu-qun ZHAO ◽  
◽  
Keyword(s):  

2014 ◽  
Vol 24 (3) ◽  
pp. 651-662
Author(s):  
Feng ZENG ◽  
Tong YANG ◽  
Shan YAO

2018 ◽  
Vol 30 (4) ◽  
pp. 642
Author(s):  
Guichao Lin ◽  
Yunchao Tang ◽  
Xiangjun Zou ◽  
Qing Zhang ◽  
Xiaojie Shi ◽  
...  

Sign in / Sign up

Export Citation Format

Share Document