Object-Oriented Building Extraction from High-Resolution Imagery Based on Fuzzy SVM

Author(s):  
Yu Meng ◽  
Shu Peng
Land ◽  
2019 ◽  
Vol 8 (12) ◽  
pp. 193
Author(s):  
Ali Alghamdi ◽  
Anthony R. Cummings

The implications of change on local processes have attracted significant research interest in recent times. In urban settings, green spaces and forests have attracted much attention. Here, we present an assessment of change within the predominantly desert Middle Eastern city of Riyadh, an understudied setting. We utilized high-resolution SPOT 5 data and two classification techniques—maximum likelihood classification and object-oriented classification—to study the changes in Riyadh between 2004 and 2014. Imagery classification was completed with training data obtained from the SPOT 5 dataset, and an accuracy assessment was completed through a combination of field surveys and an application developed in ESRI Survey 123 tool. The Survey 123 tool allowed residents of Riyadh to present their views on land cover for the 2004 and 2014 imagery. Our analysis showed that soil or ‘desert’ areas were converted to roads and buildings to accommodate for Riyadh’s rapidly growing population. The object-oriented classifier provided higher overall accuracy than the maximum likelihood classifier (74.71% and 73.79% vs. 92.36% and 90.77% for 2004 and 2014). Our work provides insights into the changes within a desert environment and establishes a foundation for understanding change in this understudied setting.


2018 ◽  
Vol 10 (11) ◽  
pp. 1768 ◽  
Author(s):  
Hui Yang ◽  
Penghai Wu ◽  
Xuedong Yao ◽  
Yanlan Wu ◽  
Biao Wang ◽  
...  

Building extraction from very high resolution (VHR) imagery plays an important role in urban planning, disaster management, navigation, updating geographic databases, and several other geospatial applications. Compared with the traditional building extraction approaches, deep learning networks have recently shown outstanding performance in this task by using both high-level and low-level feature maps. However, it is difficult to utilize different level features rationally with the present deep learning networks. To tackle this problem, a novel network based on DenseNets and the attention mechanism was proposed, called the dense-attention network (DAN). The DAN contains an encoder part and a decoder part which are separately composed of lightweight DenseNets and a spatial attention fusion module. The proposed encoder–decoder architecture can strengthen feature propagation and effectively bring higher-level feature information to suppress the low-level feature and noises. Experimental results based on public international society for photogrammetry and remote sensing (ISPRS) datasets with only red–green–blue (RGB) images demonstrated that the proposed DAN achieved a higher score (96.16% overall accuracy (OA), 92.56% F1 score, 90.56% mean intersection over union (MIOU), less training and response time and higher-quality value) when compared with other deep learning methods.


2020 ◽  
Vol 12 (22) ◽  
pp. 3764
Author(s):  
Peng Zhang ◽  
Peijun Du ◽  
Cong Lin ◽  
Xin Wang ◽  
Erzhu Li ◽  
...  

Automated extraction of buildings from earth observation (EO) data has long been a fundamental but challenging research topic. Combining data from different modalities (e.g., high-resolution imagery (HRI) and light detection and ranging (LiDAR) data) has shown great potential in building extraction. Recent studies have examined the role that deep learning (DL) could play in both multimodal data fusion and urban object extraction. However, DL-based multimodal fusion networks may encounter the following limitations: (1) the individual modal and cross-modal features, which we consider both useful and important for final prediction, cannot be sufficiently learned and utilized and (2) the multimodal features are fused by a simple summation or concatenation, which appears ambiguous in selecting cross-modal complementary information. In this paper, we address these two limitations by proposing a hybrid attention-aware fusion network (HAFNet) for building extraction. It consists of RGB-specific, digital surface model (DSM)-specific, and cross-modal streams to sufficiently learn and utilize both individual modal and cross-modal features. Furthermore, an attention-aware multimodal fusion block (Att-MFBlock) was introduced to overcome the fusion problem by adaptively selecting and combining complementary features from each modality. Extensive experiments conducted on two publicly available datasets demonstrated the effectiveness of the proposed HAFNet for building extraction.


Sign in / Sign up

Export Citation Format

Share Document