scholarly journals Recognition of Point Sets Objects in Realistic Scenes

2020 ◽  
Vol 2020 ◽  
pp. 1-13
Author(s):  
Ruizhen Gao ◽  
Xiaohui Li ◽  
Jingjun Zhang

With the emergence of new intelligent sensing technologies such as 3D scanners and stereo vision, high-quality point clouds have become very convenient and lower cost. The research of 3D object recognition based on point clouds has also received widespread attention. Point clouds are an important type of geometric data structure. Because of its irregular format, many researchers convert this data into regular three-dimensional voxel grids or image collections. However, this can lead to unnecessary bulk of data and cause problems. In this paper, we consider the problem of recognizing objects in realistic senses. We first use Euclidean distance clustering method to segment objects in realistic scenes. Then we use a deep learning network structure to directly extract features of the point cloud data to recognize the objects. Theoretically, this network structure shows strong performance. In experiment, there is an accuracy rate of 98.8% on the training set, and the accuracy rate in the experimental test set can reach 89.7%. The experimental results show that the network structure in this paper can accurately identify and classify point cloud objects in realistic scenes and maintain a certain accuracy when the number of point clouds is small, which is very robust.

Sensors ◽  
2021 ◽  
Vol 21 (3) ◽  
pp. 884
Author(s):  
Chia-Ming Tsai ◽  
Yi-Horng Lai ◽  
Yung-Da Sun ◽  
Yu-Jen Chung ◽  
Jau-Woei Perng

Numerous sensors can obtain images or point cloud data on land, however, the rapid attenuation of electromagnetic signals and the lack of light in water have been observed to restrict sensing functions. This study expands the utilization of two- and three-dimensional detection technologies in underwater applications to detect abandoned tires. A three-dimensional acoustic sensor, the BV5000, is used in this study to collect underwater point cloud data. Some pre-processing steps are proposed to remove noise and the seabed from raw data. Point clouds are then processed to obtain two data types: a 2D image and a 3D point cloud. Deep learning methods with different dimensions are used to train the models. In the two-dimensional method, the point cloud is transferred into a bird’s eye view image. The Faster R-CNN and YOLOv3 network architectures are used to detect tires. Meanwhile, in the three-dimensional method, the point cloud associated with a tire is cut out from the raw data and is used as training data. The PointNet and PointConv network architectures are then used for tire classification. The results show that both approaches provide good accuracy.


Sensors ◽  
2020 ◽  
Vol 21 (1) ◽  
pp. 201
Author(s):  
Michael Bekele Maru ◽  
Donghwan Lee ◽  
Kassahun Demissie Tola ◽  
Seunghee Park

Modeling a structure in the virtual world using three-dimensional (3D) information enhances our understanding, while also aiding in the visualization, of how a structure reacts to any disturbance. Generally, 3D point clouds are used for determining structural behavioral changes. Light detection and ranging (LiDAR) is one of the crucial ways by which a 3D point cloud dataset can be generated. Additionally, 3D cameras are commonly used to develop a point cloud containing many points on the external surface of an object around it. The main objective of this study was to compare the performance of optical sensors, namely a depth camera (DC) and terrestrial laser scanner (TLS) in estimating structural deflection. We also utilized bilateral filtering techniques, which are commonly used in image processing, on the point cloud data for enhancing their accuracy and increasing the application prospects of these sensors in structure health monitoring. The results from these sensors were validated by comparing them with the outputs from a linear variable differential transformer sensor, which was mounted on the beam during an indoor experiment. The results showed that the datasets obtained from both the sensors were acceptable for nominal deflections of 3 mm and above because the error range was less than ±10%. However, the result obtained from the TLS were better than those obtained from the DC.


Author(s):  
Y. Hori ◽  
T. Ogawa

The implementation of laser scanning in the field of archaeology provides us with an entirely new dimension in research and surveying. It allows us to digitally recreate individual objects, or entire cities, using millions of three-dimensional points grouped together in what is referred to as "point clouds". In addition, the visualization of the point cloud data, which can be used in the final report by archaeologists and architects, should usually be produced as a JPG or TIFF file. Not only the visualization of point cloud data, but also re-examination of older data and new survey of the construction of Roman building applying remote-sensing technology for precise and detailed measurements afford new information that may lead to revising drawings of ancient buildings which had been adduced as evidence without any consideration of a degree of accuracy, and finally can provide new research of ancient buildings. We used laser scanners at fields because of its speed, comprehensive coverage, accuracy and flexibility of data manipulation. Therefore, we “skipped” many of post-processing and focused on the images created from the meta-data simply aligned using a tool which extended automatic feature-matching algorithm and a popular renderer that can provide graphic results.


2013 ◽  
Vol 760-762 ◽  
pp. 1556-1561
Author(s):  
Ting Wei Du ◽  
Bo Liu

Indoor scene understanding based on the depth image data is a cutting-edge issue in the field of three-dimensional computer vision. Taking the layout characteristics of the indoor scenes and more plane features in these scenes into account, this paper presents a depth image segmentation method based on Gauss Mixture Model clustering. First, transform the Kinect depth image data into point cloud which is in the form of discrete three-dimensional point data, and denoise and down-sample the point cloud data; second, calculate the point normal of all points in the entire point cloud, then cluster the entire normal using Gaussian Mixture Model, and finally implement the entire point clouds segmentation by RANSAC algorithm. Experimental results show that the divided regions have obvious boundaries and segmentation quality is above normal, and lay a good foundation for object recognition.


2019 ◽  
Vol 8 (5) ◽  
pp. 213 ◽  
Author(s):  
Florent Poux ◽  
Roland Billen

Automation in point cloud data processing is central in knowledge discovery within decision-making systems. The definition of relevant features is often key for segmentation and classification, with automated workflows presenting the main challenges. In this paper, we propose a voxel-based feature engineering that better characterize point clusters and provide strong support to supervised or unsupervised classification. We provide different feature generalization levels to permit interoperable frameworks. First, we recommend a shape-based feature set (SF1) that only leverages the raw X, Y, Z attributes of any point cloud. Afterwards, we derive relationship and topology between voxel entities to obtain a three-dimensional (3D) structural connectivity feature set (SF2). Finally, we provide a knowledge-based decision tree to permit infrastructure-related classification. We study SF1/SF2 synergy on a new semantic segmentation framework for the constitution of a higher semantic representation of point clouds in relevant clusters. Finally, we benchmark the approach against novel and best-performing deep-learning methods while using the full S3DIS dataset. We highlight good performances, easy-integration, and high F1-score (> 85%) for planar-dominant classes that are comparable to state-of-the-art deep learning.


Author(s):  
Gülhan Benli

Since the 2000s, terrestrial laser scanning, as one of the methods used to document historical edifices in protected areas, has taken on greater importance because it mitigates the difficulties associated with working on large areas and saves time while also making it possible to better understand all the particularities of the area. Through this technology, comprehensive point data (point clouds) about the surface of an object can be generated in a highly accurate three-dimensional manner. Furthermore, with the proper software this three-dimensional point cloud data can be transformed into three-dimensional rendering/mapping/modeling and quantitative orthophotographs. In this chapter, the study will present the results of terrestrial laser scanning and surveying which was used to obtain three-dimensional point clouds through three-dimensional survey measurements and scans of silhouettes of streets in Fatih in Historic Peninsula in Istanbul, which were then transposed into survey images and drawings. The study will also cite examples of the facade mapping using terrestrial laser scanning data in Istanbul Historic Peninsula Project.


2020 ◽  
Author(s):  
Yuichi S. Hayakawa ◽  
Hiroyuki Obanawa

<p>Measuring three-dimensional morphological changes in rocky coasts is essential in protecting the </p><p>coastal areas and evaluating the sediment dynamics therein. In this study, we carried out repeated </p><p>measurements of the three-dimensional morphology of a small rocky island using terrestrial laser </p><p>scanning (TLS) and unmanned aerial vehicle (UAV)-based structure-from-motion (SfM) </p><p>photogrammetry for 5 years. The TLS-derived point cloud data is used to align the UAV-SfM point </p><p>cloud with a better accuracy at a centimeters scale, for which iterative closest point (ICP) method was </p><p>applied. Aligned UAV-derived point clouds were then compared each other to extract changed mass </p><p>for each time period. The extracted point cloud of changed mass was converted to 3D mesh polygons, </p><p>by which the total volume of eroded mass was calculated.</p><p>The temporal analysis of the point cloud revealed spatially variable rockfalls and wave cuts. The </p><p>eroded mass volume for each period varied from 10.6 to 527.7 m3, which is equivalent to the horizontal </p><p>erosion rates of 0.03 to 0.63 m/y. The temporal changes in the eroded volume is roughly associated </p><p>with that in the frequency of high tidal waves (higher than 3 m) observed in this area. However, less </p><p>correlation was found with the frequency of large ground shakes by earthquakes. The modern erosion </p><p>rate is lower than the previously reported cliff retreat rates, but this suggests that the small island will </p><p>disappear in decades. Three-dimensional structural analysis will also help understand the dynamic </p><p>processes of the erosion of the bedrock cliffs in the island.</p>


Author(s):  
R. Cura ◽  
J. Perret ◽  
N. Paparoditis

In addition to the traditional Geographic Information System (GIS) data such as images and vectors, point cloud data has become more available. It is appreciated for its precision and true three-Dimensional (3D) nature. However, managing the point cloud can be difficult due to scaling problems and specificities of this data type. Several methods exist but are usually fairly specialised and solve only one aspect of the management problem. In this work, we propose a complete and efficient point cloud management system based on a database server that works on groups of points rather than individual points. This system is specifically designed to solve all the needs of point cloud users: fast loading, compressed storage, powerful filtering, easy data access and exporting, and integrated processing. Moreover, the system fully integrates metadata (like sensor position) and can conjointly use point clouds with images, vectors, and other point clouds. The system also offers in-base processing for easy prototyping and parallel processing and can scale well. Lastly, the system is built on open source technologies; therefore it can be easily extended and customised. We test the system will several <i>billion</i> points of point clouds from Lidar (aerial and terrestrial ) and stereo-vision. We demonstrate ~ <i>400 million pts/h</i> loading speed, user-transparent greater than <i>2 to 4:1</i> compression ratio, filtering in the approximately <i>50 ms</i> range, and output of about a million pts/s, along with classical processing, such as object detection.


2021 ◽  
Author(s):  
Simone Müller ◽  
Dieter Kranzlmüller

Based on depth perception of individual stereo cameras, spatial structures can be derived as point clouds. The quality of such three-dimensional data is technically restricted by sensor limitations, latency of recording, and insufficient object reconstructions caused by surface illustration. Additionally external physical effects like lighting conditions, material properties, and reflections can lead to deviations between real and virtual object perception. Such physical influences can be seen in rendered point clouds as geometrical imaging errors on surfaces and edges. We propose the simultaneous use of multiple and dynamically arranged cameras. The increased information density leads to more details in surrounding detection and object illustration. During a pre-processing phase the collected data are merged and prepared. Subsequently, a logical analysis part examines and allocates the captured images to three-dimensional space. For this purpose, it is necessary to create a new metadata set consisting of image and localisation data. The post-processing reworks and matches the locally assigned images. As a result, the dynamic moving images become comparable so that a more accurate point cloud can be generated. For evaluation and better comparability we decided to use synthetically generated data sets. Our approach builds the foundation for dynamic and real-time based generation of digital twins with the aid of real sensor data.


Author(s):  
M. Nakagawa ◽  
R. Nozaki

<p><strong>Abstract.</strong> Three-dimensional indoor navigation requires various functions, such as the shortest path retrieval, obstacle avoidance, and secure path retrieval, for optimal path finding using a geometrical network model. Although the geometrical network model can be prepared manually, the model should be automatically generated using images and point clouds to represent changing indoor environments. Thus, we propose a methodology for generating a geometrical network model for indoor navigation using point clouds through object classification, navigable area estimation, and navigable path estimation. Our proposed methodology was evaluated through experiments using the benchmark of the International Society for Photogrammetry and Remote Sensing for indoor modeling. In our experiments, we confirmed that our methodology can generate a geometrical network model automatically.</p>


Sign in / Sign up

Export Citation Format

Share Document