scholarly journals Correcting Image Refraction: Towards Accurate Aerial Image-Based Bathymetry Mapping in Shallow Waters

2020 ◽  
Vol 12 (2) ◽  
pp. 322 ◽  
Author(s):  
Panagiotis Agrafiotis ◽  
Konstantinos Karantzalos ◽  
Andreas Georgopoulos ◽  
Dimitrios Skarlatos

Although aerial image-based bathymetric mapping can provide, unlike acoustic or LiDAR (Light Detection and Ranging) sensors, both water depth and visual information, water refraction poses significant challenges for accurate depth estimation. In order to tackle this challenge, we propose an image correction methodology, which first exploits recent machine learning procedures that recover depth from image-based dense point clouds and then corrects refraction on the original imaging dataset. This way, the structure from motion (SfM) and multi-view stereo (MVS) processing pipelines are executed on a refraction-free set of aerial datasets, resulting in highly accurate bathymetric maps. Performed experiments and validation were based on datasets acquired during optimal sea state conditions and derived from four different test-sites characterized by excellent sea bottom visibility and textured seabed. Results demonstrated the high potential of our approach, both in terms of bathymetric accuracy, as well as texture and orthoimage quality.

Author(s):  
Guillermo Oliver ◽  
Pablo Gil ◽  
Jose F. Gomez ◽  
Fernando Torres

AbstractIn this paper, we present a robotic workcell for task automation in footwear manufacturing such as sole digitization, glue dispensing, and sole manipulation from different places within the factory plant. We aim to make progress towards shoe industry 4.0. To achieve it, we have implemented a novel sole grasping method, compatible with soles of different shapes, sizes, and materials, by exploiting the particular characteristics of these objects. Our proposal is able to work well with low density point clouds from a single RGBD camera and also with dense point clouds obtained from a laser scanner digitizer. The method computes antipodal grasping points from visual data in both cases and it does not require a previous recognition of sole. It relies on sole contour extraction using concave hulls and measuring the curvature on contour areas. Our method was tested both in a simulated environment and in real conditions of manufacturing at INESCOP facilities, processing 20 soles with different sizes and characteristics. Grasps were performed in two different configurations, obtaining an average score of 97.5% of successful real grasps for soles without heel made with materials of low or medium flexibility. In both cases, the grasping method was tested without carrying out tactile control throughout the task.


2019 ◽  
Vol 93 (3) ◽  
pp. 411-429 ◽  
Author(s):  
Maria Immacolata Marzulli ◽  
Pasi Raumonen ◽  
Roberto Greco ◽  
Manuela Persia ◽  
Patrizia Tartarino

Abstract Methods for the three-dimensional (3D) reconstruction of forest trees have been suggested for data from active and passive sensors. Laser scanner technologies have become popular in the last few years, despite their high costs. Since the improvements in photogrammetric algorithms (e.g. structure from motion—SfM), photographs have become a new low-cost source of 3D point clouds. In this study, we use images captured by a smartphone camera to calculate dense point clouds of a forest plot using SfM. Eighteen point clouds were produced by changing the densification parameters (Image scale, Point density, Minimum number of matches) in order to investigate their influence on the quality of the point clouds produced. In order to estimate diameter at breast height (d.b.h.) and stem volumes, we developed an automatic method that extracts the stems from the point cloud and then models them with cylinders. The results show that Image scale is the most influential parameter in terms of identifying and extracting trees from the point clouds. The best performance with cylinder modelling from point clouds compared to field data had an RMSE of 1.9 cm and 0.094 m3, for d.b.h. and volume, respectively. Thus, for forest management and planning purposes, it is possible to use our photogrammetric and modelling methods to measure d.b.h., stem volume and possibly other forest inventory metrics, rapidly and without felling trees. The proposed methodology significantly reduces working time in the field, using ‘non-professional’ instruments and automating estimates of dendrometric parameters.


Author(s):  
L. Madhuanand ◽  
F. Nex ◽  
M. Y. Yang

Abstract. Depth is an essential component for various scene understanding tasks and for reconstructing the 3D geometry of the scene. Estimating depth from stereo images requires multiple views of the same scene to be captured which is often not possible when exploring new environments with a UAV. To overcome this monocular depth estimation has been a topic of interest with the recent advancements in computer vision and deep learning techniques. This research has been widely focused on indoor scenes or outdoor scenes captured at ground level. Single image depth estimation from aerial images has been limited due to additional complexities arising from increased camera distance, wider area coverage with lots of occlusions. A new aerial image dataset is prepared specifically for this purpose combining Unmanned Aerial Vehicles (UAV) images covering different regions, features and point of views. The single image depth estimation is based on image reconstruction techniques which uses stereo images for learning to estimate depth from single images. Among the various available models for ground-level single image depth estimation, two models, 1) a Convolutional Neural Network (CNN) and 2) a Generative Adversarial model (GAN) are used to learn depth from aerial images from UAVs. These models generate pixel-wise disparity images which could be converted into depth information. The generated disparity maps from these models are evaluated for its internal quality using various error metrics. The results show higher disparity ranges with smoother images generated by CNN model and sharper images with lesser disparity range generated by GAN model. The produced disparity images are converted to depth information and compared with point clouds obtained using Pix4D. It is found that the CNN model performs better than GAN and produces depth similar to that of Pix4D. This comparison helps in streamlining the efforts to produce depth from a single aerial image.


Author(s):  
R. Moritani ◽  
S. Kanai ◽  
H. Date ◽  
Y. Niina ◽  
R. Honma

<p><strong>Abstract.</strong> In this paper, we introduce a method for predicting the quality of dense points and selecting low-quality regions on the points generated by the structure from motion (SfM) and multi-view stereo (MVS) pipeline to realize high-quality and efficient as-is model reconstruction, using only results from the former: sparse point clouds and camera poses. The method was shown to estimate the quality of the final dense points as the quality predictor on an approximated model obtained from SfM only, without requiring the time-consuming MVS process. Moreover, the predictors can be used for selection of low-quality regions on the approximated model to estimate the next-best optimum camera poses which could improve quality. Furthermore, the method was applied to the prediction of dense point quality generated from the image sets of a concrete bridge column and construction site, and the prediction was validated in a time much shorter than using MVS. Finally, we discussed the correlation between the predictors and the final dense point quality.</p>


Sensors ◽  
2019 ◽  
Vol 19 (3) ◽  
pp. 563 ◽  
Author(s):  
J. Osuna-Coutiño ◽  
Jose Martinez-Carranza

High-Level Structure (HLS) extraction in a set of images consists of recognizing 3D elements with useful information to the user or application. There are several approaches to HLS extraction. However, most of these approaches are based on processing two or more images captured from different camera views or on processing 3D data in the form of point clouds extracted from the camera images. In contrast and motivated by the extensive work developed for the problem of depth estimation in a single image, where parallax constraints are not required, in this work, we propose a novel methodology towards HLS extraction from a single image with promising results. For that, our method has four steps. First, we use a CNN to predict the depth for a single image. Second, we propose a region-wise analysis to refine depth estimates. Third, we introduce a graph analysis to segment the depth in semantic orientations aiming at identifying potential HLS. Finally, the depth sections are provided to a new CNN architecture that predicts HLS in the shape of cubes and rectangular parallelepipeds.


Author(s):  
David Lee ◽  
William Muir ◽  
Samuel Beeston ◽  
Samuel Bates ◽  
Sam D. Schofield ◽  
...  
Keyword(s):  

2019 ◽  
Vol 11 (16) ◽  
pp. 1940 ◽  
Author(s):  
Fausto Mistretta ◽  
Giannina Sanna ◽  
Flavio Stochino ◽  
Giuseppina Vacca

Dense point clouds acquired from Terrestrial Laser Scanners (TLS) have proved to be effective for structural deformation assessment. In the last decade, many researchers have defined methodology and workflow in order to compare different point clouds, with respect to each other or to a known model, assessing the potentialities and limits of this technique. Currently, dense point clouds can be obtained by Close-Range Photogrammetry (CRP) based on a Structure from Motion (SfM) algorithm. This work reports on a comparison between the TLS technique and the Close-Range Photogrammetry using the Structure from Motion algorithm. The analysis of two Reinforced Concrete (RC) beams tested under four-points bending loading is presented. In order to measure displacement distributions, point clouds at different beam loading states were acquired and compared. A description of the instrumentation used and the experimental environment, along with a comprehensive report on the calculations and results obtained is reported. Two kinds of point clouds comparison were investigated: Mesh to mesh and modeling with geometric primitives. The comparison between the mesh to mesh (m2m) approach and the modeling (m) one showed that the latter leads to significantly better results for both TLS and CRP. The results obtained with the TLS for both m2m and m methodologies present a Root Mean Square (RMS) levels below 1 mm, while the CRP method yields to an RMS level of a few millimeters for m2m, and of 1 mm for m.


Sign in / Sign up

Export Citation Format

Share Document