Fully‐connected semantic segmentation of hyperspectral and LiDAR data

2019 ◽  
Vol 13 (3) ◽  
pp. 285-293
Author(s):  
Hakan Aytaylan ◽  
Seniha Esen Yuksel
Sensors ◽  
2020 ◽  
Vol 20 (12) ◽  
pp. 3568 ◽  
Author(s):  
Takayuki Shinohara ◽  
Haoyi Xiu ◽  
Masashi Matsuoka

In the computer vision field, many 3D deep learning models that directly manage 3D point clouds (proposed after PointNet) have been published. Moreover, deep learning-based-techniques have demonstrated state-of-the-art performance for supervised learning tasks on 3D point cloud data, such as classification and segmentation tasks for open datasets in competitions. Furthermore, many researchers have attempted to apply these deep learning-based techniques to 3D point clouds observed by aerial laser scanners (ALSs). However, most of these studies were developed for 3D point clouds without radiometric information. In this paper, we investigate the possibility of using a deep learning method to solve the semantic segmentation task of airborne full-waveform light detection and ranging (lidar) data that consists of geometric information and radiometric waveform data. Thus, we propose a data-driven semantic segmentation model called the full-waveform network (FWNet), which handles the waveform of full-waveform lidar data without any conversion process, such as projection onto a 2D grid or calculating handcrafted features. Our FWNet is based on a PointNet-based architecture, which can extract the local and global features of each input waveform data, along with its corresponding geographical coordinates. Subsequently, the classifier consists of 1D convolutional operational layers, which predict the class vector corresponding to the input waveform from the extracted local and global features. Our trained FWNet achieved higher scores in its recall, precision, and F1 score for unseen test data—higher scores than those of previously proposed methods in full-waveform lidar data analysis domain. Specifically, our FWNet achieved a mean recall of 0.73, a mean precision of 0.81, and a mean F1 score of 0.76. We further performed an ablation study, that is assessing the effectiveness of our proposed method, of the above-mentioned metric. Moreover, we investigated the effectiveness of our PointNet based local and global feature extraction method using the visualization of the feature vector. In this way, we have shown that our network for local and global feature extraction allows training with semantic segmentation without requiring expert knowledge on full-waveform lidar data or translation into 2D images or voxels.


Author(s):  
Weihao Li ◽  
Michael Ying Yang

In this paper we explore semantic segmentation of man-made scenes using fully connected conditional random field (CRF). Images of man-made scenes display strong contextual dependencies in the spatial structures. Fully connected CRFs can model long-range connections within the image of man-made scenes and make use of contextual information of scene structures. The pairwise edge potentials of fully connected CRF models are defined by a linear combination of Gaussian kernels. Using filter-based mean field algorithm, the inference is very efficient. Our experimental results demonstrate that fully connected CRF performs better than previous state-of-the-art approaches on both eTRIMS dataset and LabelMeFacade dataset.


Author(s):  
E. Tusa ◽  
J. M. Monnet ◽  
J. B. Barré ◽  
M. Dalla Mura ◽  
J. Chanussot

Abstract. Hyperspectral images (HI) and Light Detection and Ranging (LiDAR) provide high resolution radiometric and geometric information for monitoring forests at individual tree crown (ITC) level. It has many important applications for sustainable forest management, biodiversity assessment and healthy ecosystem preservation. However, the integration of different remote sensing modalities is a challenging task for tree species classification due to different artifacts such as the lighting variability, the topographic effects and the atmospheric conditions of the data acquisition. The characterization of ITC can benefit from the extraction and selection of robust feature descriptors that solve these issues. This paper aims to investigate the integration of feature descriptors from HI and LiDAR by using the intra-set and inter-set feature importance for the semantic segmentation of forest tree species. A fusion methodology is proposed between high-density LiDAR data – (20 pulses m−2) and VNIR HI – (160 bands and 0.80 m spatial resolution) acquired on French temperate forests along an altitude gradient. The proposed scheme has three inputs: the field inventory information, the HI and the LiDAR data. Our approach can be described in nine stages: polygon projection, non-overlapping pixel selection, vegetation and shadow removal, LiDAR feature extraction, height mask, robust PCA (rPCA), feature reduction and classification. The overall accuracy of tree species classification at pixel-level was 68.9% by using random forest (RF) classifier. Our approach showed that 74.0% of trees were correctly assigned overall, by having conifer species such as Norway Spruce (Picea abies) with a producer’s accuracy of 97.4%.


Author(s):  
H. Wu ◽  
Z. Xie ◽  
C. Wen ◽  
C. Wang ◽  
J. Li

Abstract. On-road information, including road boundaries, road markings, and road cracks, provides significant guidance or warning to all road users. Recently, the on-road information extraction from LiDAR data have been widely studied. However, for the LiDAR data with lower accuracy and higher noise, some detailed information, such as road boundary, is difficult to be extracted correctly. Furthermore, most of previous studies lack an exploration of efficiently extracting multiple on-road information from a single framework. In this paper, we propose a new framework that can simultaneously extract multiple on-road information from high accuracy LiDAR data and can also more robustly extract detailed road boundaries from low accuracy LiDAR data. First, we propose a Curb-Aware Ground Filter to extract ground points with rich curb structure features. Second, we transform the vertical density, elevation gradient and intensity features of the ground points into multiple feature maps and extract multiple on-road information from the feature maps by employing a semantic segmentation network. Experimental results on three datasets with different data accuracy demonstrate that our method outperforms other recent competitive methods.


Author(s):  
B. Vishnyakov ◽  
Y. Blokhinov ◽  
I. Sgibnev ◽  
V. Sheverdin ◽  
A. Sorokin ◽  
...  

Abstract. In this paper we describe a new multi-sensor platform for data collection and algorithm testing. We propose a couple of methods for solution of semantic scene understanding problem for land autonomous vehicles. We describe our approaches for automatic camera and LiDAR calibration; three-dimensional scene reconstruction and odometry calculation; semantic segmentation that provides obstacle recognition and underlying surface classification; object detection; point cloud segmentation. Also, we describe our virtual simulation complex based on Unreal Engine, that can be used for both data collection and algorithm testing. We collected a large database of field and virtual data: more than 1,000,000 real images with corresponding LiDAR data and more than 3,500,000 simulated images with corresponding LiDAR data. All proposed methods were implemented and tested on our autonomous platform; accuracy estimates were obtained on the collected database.


Sign in / Sign up

Export Citation Format

Share Document