scholarly journals Correction to ``Scene-Driven Multitask Parallel Attention Network for Building Extraction in High-Resolution Remote Sensing Images''

Author(s):  
Haonan Guo ◽  
Qian Shi ◽  
Bo Du ◽  
Liangpei Zhang ◽  
Dongzhi Wang ◽  
...  
2020 ◽  
Vol 12 (6) ◽  
pp. 1050 ◽  
Author(s):  
Zhenfeng Shao ◽  
Penghao Tang ◽  
Zhongyuan Wang ◽  
Nayyer Saleem ◽  
Sarath Yam ◽  
...  

Building extraction from high-resolution remote sensing images is of great significance in urban planning, population statistics, and economic forecast. However, automatic building extraction from high-resolution remote sensing images remains challenging. On the one hand, the extraction results of buildings are partially missing and incomplete due to the variation of hue and texture within a building, especially when the building size is large. On the other hand, the building footprint extraction of buildings with complex shapes is often inaccurate. To this end, we propose a new deep learning network, termed Building Residual Refine Network (BRRNet), for accurate and complete building extraction. BRRNet consists of such two parts as the prediction module and the residual refinement module. The prediction module based on an encoder–decoder structure introduces atrous convolution of different dilation rates to extract more global features, by gradually increasing the receptive field during feature extraction. When the prediction module outputs the preliminary building extraction results of the input image, the residual refinement module takes the output of the prediction module as an input. It further refines the residual between the result of the prediction module and the real result, thus improving the accuracy of building extraction. In addition, we use Dice loss as the loss function during training, which effectively alleviates the problem of data imbalance and further improves the accuracy of building extraction. The experimental results on Massachusetts Building Dataset show that our method outperforms other five state-of-the-art methods in terms of the integrity of buildings and the accuracy of complex building footprints.


2020 ◽  
Vol 86 (4) ◽  
pp. 235-245 ◽  
Author(s):  
Ka Zhang ◽  
Hui Chen ◽  
Wen Xiao ◽  
Yehua Sheng ◽  
Dong Su ◽  
...  

This article proposes a new building extraction method from high-resolution remote sensing images, based on GrabCut, which can automatically select foreground and background samples under the constraints of building elevation contour lines. First the image is rotated according to the direction of pixel displacement calculated by the rational function Model. Second, the Canny operator, combined with morphology and the Hough transform, is used to extract the building's elevation contour lines. Third, seed points and interesting points of the building are selected under the constraint of the contour line and the geodesic distance. Then foreground and background samples are obtained according to these points. Fourth, GrabCut and geometric features are used to carry out image segmentation and extract buildings. Finally, WorldView satellite images are used to verify the proposed method. Experimental results show that the average accuracy can reach 86.34%, which is 15.12% higher than other building extraction methods.


Sensors ◽  
2020 ◽  
Vol 20 (24) ◽  
pp. 7241
Author(s):  
Dengji Zhou ◽  
Guizhou Wang ◽  
Guojin He ◽  
Tengfei Long ◽  
Ranyu Yin ◽  
...  

Building extraction from high spatial resolution remote sensing images is a hot spot in the field of remote sensing applications and computer vision. This paper presents a semantic segmentation model, which is a supervised method, named Pyramid Self-Attention Network (PISANet). Its structure is simple, because it contains only two parts: one is the backbone of the network, which is used to learn the local features (short distance context information around the pixel) of buildings from the image; the other part is the pyramid self-attention module, which is used to obtain the global features (long distance context information with other pixels in the image) and the comprehensive features (includes color, texture, geometric and high-level semantic feature) of the building. The network is an end-to-end approach. In the training stage, the input is the remote sensing image and corresponding label, and the output is probability map (the probability that each pixel is or is not building). In the prediction stage, the input is the remote sensing image, and the output is the extraction result of the building. The complexity of the network structure was reduced so that it is easy to implement. The proposed PISANet was tested on two datasets. The result shows that the overall accuracy reached 94.50 and 96.15%, the intersection-over-union reached 77.45 and 87.97%, and F1 index reached 87.27 and 93.55%, respectively. In experiments on different datasets, PISANet obtained high overall accuracy, low error rate and improved integrity of individual buildings.


2019 ◽  
Vol 11 (23) ◽  
pp. 2813 ◽  
Author(s):  
Wenchao Kang ◽  
Yuming Xiang ◽  
Feng Wang ◽  
Hongjian You

Automatic building extraction from high-resolution remote sensing images has many practical applications, such as urban planning and supervision. However, fine details and various scales of building structures in high-resolution images bring new challenges to building extraction. An increasing number of neural network-based models have been proposed to handle these issues, while they are not efficient enough, and still suffer from the error ground truth labels. To this end, we propose an efficient end-to-end model, EU-Net, in this paper. We first design the dense spatial pyramid pooling (DSPP) to extract dense and multi-scale features simultaneously, which facilitate the extraction of buildings at all scales. Then, the focal loss is used in reverse to suppress the impact of the error labels in ground truth, making the training stage more stable. To assess the universality of the proposed model, we tested it on three public aerial remote sensing datasets: WHU aerial imagery dataset, Massachusetts buildings dataset, and Inria aerial image labeling dataset. Experimental results show that the proposed EU-Net is superior to the state-of-the-art models of all three datasets and increases the prediction efficiency by two to four times.


2022 ◽  
Vol 14 (2) ◽  
pp. 269
Author(s):  
Yong Wang ◽  
Xiangqiang Zeng ◽  
Xiaohan Liao ◽  
Dafang Zhuang

Deep learning (DL) shows remarkable performance in extracting buildings from high resolution remote sensing images. However, how to improve the performance of DL based methods, especially the perception of spatial information, is worth further study. For this purpose, we proposed a building extraction network with feature highlighting, global awareness, and cross level information fusion (B-FGC-Net). The residual learning and spatial attention unit are introduced in the encoder of the B-FGC-Net, which simplifies the training of deep convolutional neural networks and highlights the spatial information representation of features. The global feature information awareness module is added to capture multiscale contextual information and integrate the global semantic information. The cross level feature recalibration module is used to bridge the semantic gap between low and high level features to complete the effective fusion of cross level information. The performance of the proposed method was tested on two public building datasets and compared with classical methods, such as UNet, LinkNet, and SegNet. Experimental results demonstrate that B-FGC-Net exhibits improved profitability of accurate extraction and information integration for both small and large scale buildings. The IoU scores of B-FGC-Net on WHU and INRIA Building datasets are 90.04% and 79.31%, respectively. B-FGC-Net is an effective and recommended method for extracting buildings from high resolution remote sensing images.


Sign in / Sign up

Export Citation Format

Share Document