An advanced method for matching partial 3D point clouds to free-form CAD models for in-situ inspection and repair

Author(s):  
Bilal Nasser ◽  
Amir Rabani
Author(s):  
M. Karpina ◽  
M. Jarząbek-Rychard ◽  
P. Tymków ◽  
A. Borkowski

Manual in-situ measurements of geometric tree parameters for the biomass volume estimation are time-consuming and economically non-effective. Photogrammetric techniques can be deployed in order to automate the measurement procedure. The purpose of the presented work is an automatic tree growth estimation based on Unmanned Aircraft Vehicle (UAV) imagery. The experiment was conducted in an agriculture test field with scots pine canopies. The data was collected using a Leica Aibotix X6V2 platform equipped with a Nikon D800 camera. Reference geometric parameters of selected sample plants were measured manually each week. In situ measurements were correlated with the UAV data acquisition. The correlation aimed at the investigation of optimal conditions for a flight and parameter settings for image acquisition. The collected images are processed in a state of the art tool resulting in a generation of dense 3D point clouds. The algorithm is developed in order to estimate geometric tree parameters from 3D points. Stem positions and tree tops are identified automatically in a cross section, followed by the calculation of tree heights. The automatically derived height values are compared to the reference measurements performed manually. The comparison allows for the evaluation of automatic growth estimation process. The accuracy achieved using UAV photogrammetry for tree heights estimation is about 5cm.


2020 ◽  
Vol 9 (12) ◽  
pp. 759
Author(s):  
Yufu Zang ◽  
Bijun Li ◽  
Xiongwu Xiao ◽  
Jianfeng Zhu ◽  
Fancong Meng

Heritage documentation is implemented by digitally recording historical artifacts for the conservation and protection of these cultural heritage objects. As efficient spatial data acquisition tools, laser scanners have been widely used to collect highly accurate three-dimensional (3D) point clouds without damaging the original structure and the environment. To ensure the integrity and quality of the collected data, field inspection (i.e., on-spot checking the data quality) should be carried out to determine the need for additional measurements (i.e., extra laser scanning for areas with quality issues such as data missing and quality degradation). To facilitate inspection of all collected point clouds, especially checking the quality issues in overlaps between adjacent scans, all scans should be registered together. Thus, a point cloud registration method that is able to register scans fast and robustly is required. To fulfill the aim, this study proposes an efficient probabilistic registration for free-form cultural heritage objects by integrating the proposed principal direction descriptor and curve constraints. We developed a novel shape descriptor based on a local frame of principal directions. Within the frame, its density and distance feature images were generated to describe the shape of the local surface. We then embedded the descriptor into a probabilistic framework to reject ambiguous matches. Spatial curves were integrated as constraints to delimit the solution space. Finally, a multi-view registration was used to refine the position and orientation of each scan for the field inspection. Comprehensive experiments show that the proposed method was able to perform well in terms of rotation error, translation error, robustness, and runtime and outperformed some commonly used approaches.


Author(s):  
C. Jepping ◽  
F. Bethmann ◽  
T. Luhmann

This paper deals with the correction of exterior orientation parameters of stereo image sequences over deformed free-form surfaces without control points. Such imaging situation can occur, for example, during photogrammetric car crash test recordings where onboard high-speed stereo cameras are used to measure 3D surfaces. As a result of such measurements 3D point clouds of deformed surfaces are generated for a complete stereo sequence. The first objective of this research focusses on the development and investigation of methods for the detection of corresponding spatial and temporal tie points within the stereo image sequences (by stereo image matching and 3D point tracking) that are robust enough for a reliable handling of occlusions and other disturbances that may occur. The second objective of this research is the analysis of object deformations in order to detect stable areas (congruence analysis). For this purpose a RANSAC-based method for congruence analysis has been developed. This process is based on the sequential transformation of randomly selected point groups from one epoch to another by using a 3D similarity transformation. The paper gives a detailed description of the congruence analysis. The approach has been tested successfully on synthetic and real image data.


2007 ◽  
Vol 4 (5) ◽  
pp. 629-638 ◽  
Author(s):  
Cheuk Yiu Ip ◽  
Satyandra K. Gupta

Author(s):  
M. Karpina ◽  
M. Jarząbek-Rychard ◽  
P. Tymków ◽  
A. Borkowski

Manual in-situ measurements of geometric tree parameters for the biomass volume estimation are time-consuming and economically non-effective. Photogrammetric techniques can be deployed in order to automate the measurement procedure. The purpose of the presented work is an automatic tree growth estimation based on Unmanned Aircraft Vehicle (UAV) imagery. The experiment was conducted in an agriculture test field with scots pine canopies. The data was collected using a Leica Aibotix X6V2 platform equipped with a Nikon D800 camera. Reference geometric parameters of selected sample plants were measured manually each week. In situ measurements were correlated with the UAV data acquisition. The correlation aimed at the investigation of optimal conditions for a flight and parameter settings for image acquisition. The collected images are processed in a state of the art tool resulting in a generation of dense 3D point clouds. The algorithm is developed in order to estimate geometric tree parameters from 3D points. Stem positions and tree tops are identified automatically in a cross section, followed by the calculation of tree heights. The automatically derived height values are compared to the reference measurements performed manually. The comparison allows for the evaluation of automatic growth estimation process. The accuracy achieved using UAV photogrammetry for tree heights estimation is about 5cm.


2019 ◽  
Vol 3 (2) ◽  
pp. 40 ◽  
Author(s):  
Ulrike Wissen Hayek ◽  
Kilian Müller ◽  
Fabian Göbel ◽  
Peter Kiefer ◽  
Reto Spielhofer ◽  
...  

The perception of the visual landscape impact is a significant factor explaining the public’s acceptance of energy infrastructure developments. Yet, there is lack of knowledge how people perceive and accept power lines in certain landscape types and in combination with wind turbines, a required setting to achieve goals of the energy turnaround. The goal of this work was to demonstrate how 3D point cloud visualizations could be used for an eye tracking study to systematically investigate the perception of landscape scenarios with power lines. 3D visualizations of near-natural and urban landscapes were prepared based on data from airborne and terrestrial laser scanning. These scenes were altered with varying amounts of the respective infrastructure, and they provided the stimuli in a laboratory experiment with 49 participants. Eye tracking and questionnaires served for measuring the participants’ responses. The results show that the point cloud-based simulations offered suitable stimuli for the eye tracking study. Particularly for the analysis of guided perceptions, the approach fostered an understanding of disturbing landscape elements. A comparative in situ eye tracking study is recommended to further evaluate the quality of the point cloud simulations, whether they produce similar responses as in the real world.


Author(s):  
F. Thiel ◽  
S. Discher ◽  
R. Richter ◽  
J. Döllner

<p><strong>Abstract.</strong> Emerging virtual reality (VR) technology allows immersively exploring digital 3D content on standard consumer hardware. Using in-situ or remote sensing technology, such content can be automatically derived from real-world sites. External memory algorithms allow for the non-immersive exploration of the resulting 3D point clouds on a diverse set of devices with vastly different rendering capabilities. Applications for VR environments raise additional challenges for those algorithms as they are highly sensitive towards visual artifacts that are typical for point cloud depictions (i.e., overdraw and underdraw), while simultaneously requiring higher frame rates (i.e., around 90<span class="thinspace"></span>fps instead of 30&amp;ndash;60<span class="thinspace"></span>fps). We present a rendering system for the immersive exploration and inspection of massive 3D point clouds on state-of-the-art VR devices. Based on a multi-pass rendering pipeline, we combine point-based and image-based rendering techniques to simultaneously improve the rendering performance and the visual quality. A set of interaction and locomotion techniques allows users to inspect a 3D point cloud in detail, for example by measuring distances and areas or by scaling and rotating visualized data sets. All rendering, interaction and locomotion techniques can be selected and configured dynamically, allowing to adapt the rendering system to different use cases. Tests on data sets with up to 2.6 billion points show the feasibility and scalability of our approach.</p>


Author(s):  
A. Alizadeh Naeini ◽  
A. Ahmad ◽  
M. M. Sheikholeslami ◽  
P. Claudio ◽  
G. Sohn

Abstract. Thanks to the proliferation of commodity 3D devices such as HoloLens, one can have easy access to the 3D model of indoor building objects. However, this model does not match 2D available computer-aided design (CAD) models as the as-built model. To address this problem, in this study, a 3-step registration method is proposed. First, binary images, including walls and background, are generated for the 3D point cloud (PC) and the 2D CAD model. Then, 2D-to-2D corresponding pixels (CPs) are extracted based on the intersection of walls in each binary image of PC (BIPC) and binary CAD (BCAD) model. Since the 3D PC space coordinates (XYZ) of all BIPC's pixels are known, BIPC part of the 2D-to-2D CPs can be considered 3D. Lastly, the parameters of the 8-parameter affine are estimated using the 2D-to-3D CPs, which are pixel coordinates in BCAD model as well as their correspondences in the 3D PC space. Experimental results indicate the efficiency of our proposed method compared to manual registration.


Sign in / Sign up

Export Citation Format

Share Document