scholarly journals FSRSS-Net: High-Resolution Mapping of Buildings from Middle-Resolution Satellite Images Using a Super-Resolution Semantic Segmentation Network

2021 ◽  
Vol 13 (12) ◽  
pp. 2290
Author(s):  
Tao Zhang ◽  
Hong Tang ◽  
Yi Ding ◽  
Penglong Li ◽  
Chao Ji ◽  
...  

Satellite mapping of buildings and built-up areas used to be delineated from high spatial resolution (e.g., meters or sub-meters) and middle spatial resolution (e.g., tens of meters or hundreds of meters) satellite images, respectively. To the best of our knowledge, it is important to explore a deep-learning approach to delineate high-resolution semantic maps of buildings from middle-resolution satellite images. The approach is termed as super-resolution semantic segmentation in this paper. Specifically, we design a neural network with integrated low-level image features of super-resolution and high-level semantic features of super-resolution, which is trained with Sentinel-2A images (i.e., 10 m) and higher-resolution semantic maps (i.e., 2.5 m). The network, based on super-resolution semantic segmentation features is called FSRSS-Net. In China, the 35 cities are partitioned into three groups, i.e., 19 cities for model training, four cities for quantitative testing and the other 12 cities for qualitative generalization ability analysis of the learned networks. A large-scale sample dataset is created and utilized to train and validate the performance of the FSRSS-Net, which includes 8597 training samples and 766 quantitative accuracy evaluation samples. Quantitative evaluation results show that: (1) based on the 10 m Sentinel-2A image, the FSRSS-Net can achieve super-resolution semantic segmentation and produce 2.5 m building recognition results, and there is little difference between the accuracy of 2.5 m results by FSRSS-Net and 10 m results by U-Net. More importantly, the 2.5 m building recognition results by FSRSS-Net have higher accuracy than the 2.5 m results by U-Net 10 m building recognition results interpolation up-sampling; (2) from the spatial visualization of the results, the building recognition results of 2.5 m are more precise than those of 10 m, and the outline of the building is better depicted. Qualitative analysis shows that: (1) the learned FSRSS-Net can be also well generalized to other cities that are far from training regions; (2) the FSRSS-Net can still achieve comparable results to the U-Net 2 m building recognition results, even when the U-Net is directly trained using both 2-meter resolution GF2 satellite images and corresponding semantic labels.

Information ◽  
2021 ◽  
Vol 12 (6) ◽  
pp. 230
Author(s):  
Sultan Daud Khan ◽  
Louai Alarabi ◽  
Saleh Basalamah

Land cover semantic segmentation in high-spatial resolution satellite images plays a vital role in efficient management of land resources, smart agriculture, yield estimation and urban planning. With the recent advancement in remote sensing technologies, such as satellites, drones, UAVs, and airborne vehicles, a large number of high-resolution satellite images are readily available. However, these high-resolution satellite images are complex due to increased spatial resolution and data disruption caused by different factors involved in the acquisition process. Due to these challenges, an efficient land-cover semantic segmentation model is difficult to design and develop. In this paper, we develop a hybrid deep learning model that combines the benefits of two deep models, i.e., DenseNet and U-Net. This is carried out to obtain a pixel-wise classification of land cover. The contraction path of U-Net is replaced with DenseNet to extract features of multiple scales, while long-range connections of U-Net concatenate encoder and decoder paths are used to preserve low-level features. We evaluate the proposed hybrid network on a challenging, publicly available benchmark dataset. From the experimental results, we demonstrate that the proposed hybrid network exhibits a state-of-the-art performance and beats other existing models by a considerable margin.


2019 ◽  
Author(s):  
Sawyer Reid stippa ◽  
George Petropoulos ◽  
Leonidas Toulios ◽  
Prashant K. Srivastava

Archaeological site mapping is important for both understanding the history as well as protecting them from excavation during the developmental activities. As archaeological sites generally spread over a large area, use of high spatial resolution remote sensing imagery is becoming increasingly applicable in the world. The main objective of this study was to map the land cover of the Itanos area of Crete and of its changes, with specific focus on the detection of the landscape’s archaeological features. Six satellite images were acquired from the Pleiades and WorldView-2 satellites over a period of 3 years. In addition, digital photography of two known archaeological sites was used for validation. An Object Based Image Analysis (OBIA) classification was subsequently developed using the five acquired satellite images. Two rule-sets were created, one using the standard four bands which both satellites have and another for the two WorldView-2 images their four extra bands included. Validation of the thematic maps produced from the classification scenarios confirmed a difference in accuracy amongst the five images. Comparing the results of a 4-band rule-set versus the 8-band showed a slight increase in classification accuracy using extra bands. The resultant classifications showed a good level of accuracy exceeding 70%. Yet, separating the archaeological sites from the open spaces with little or no vegetation proved challenging. This was mainly due to the high spectral similarity between rocks and the archaeological ruins. The satellite data spatial resolution allowed for the accuracy in defining larger archaeological sites, but still was a difficulty in distinguishing smaller areas of interest. The digital photography data provided a very good 3D representation for the archaeological sites, assisting as well in validating the satellite-derived classification maps. All in all, our study provided further evidence that use of high resolution imagery may allow for archaeological sites to be located, but only where they are of a suitable size archaeological features.


2021 ◽  
Vol 13 (10) ◽  
pp. 1944
Author(s):  
Xiaoming Liu ◽  
Menghua Wang

The Visible Infrared Imaging Radiometer Suite (VIIRS) onboard the Suomi National Polar-orbiting Partnership (SNPP) satellite has been a reliable source of ocean color data products, including five moderate (M) bands and one imagery (I) band normalized water-leaving radiance spectra nLw(λ). The spatial resolutions of the M-band and I-band nLw(λ) are 750 m and 375 m, respectively. With the technique of convolutional neural network (CNN), the M-band nLw(λ) imagery can be super-resolved from 750 m to 375 m spatial resolution by leveraging the high spatial resolution features of I1-band nLw(λ) data. However, it is also important to enhance the spatial resolution of VIIRS-derived chlorophyll-a (Chl-a) concentration and the water diffuse attenuation coefficient at the wavelength of 490 nm (Kd(490)), as well as other biological and biogeochemical products. In this study, we describe our effort to derive high-resolution Kd(490) and Chl-a data based on super-resolved nLw(λ) images at the VIIRS five M-bands. To improve the network performance over extremely turbid coastal oceans and inland waters, the networks are retrained with a training dataset including ocean color data from the Bohai Sea, Baltic Sea, and La Plata River Estuary, covering water types from clear open oceans to moderately turbid and highly turbid waters. The evaluation results show that the super-resolved Kd(490) image is much sharper than the original one, and has more detailed fine spatial structures. A similar enhancement of finer structures is also found in the super-resolved Chl-a images. Chl-a filaments are much sharper and thinner in the super-resolved image, and some of the very fine spatial features that are not shown in the original images appear in the super-resolved Chl-a imageries. The networks are also applied to four other coastal and inland water regions. The results show that super-resolution occurs mainly on pixels of Chl-a and Kd(490) features, especially on the feature edges and locations with a large spatial gradient. The biases between the original M-band images and super-resolved high-resolution images are small for both Chl-a and Kd(490) in moderately to extremely turbid coastal oceans and inland waters, indicating that the super-resolution process does not change the mean values of the original images.


Author(s):  
R. S. Hansen ◽  
D. W. Waldram ◽  
T. Q. Thai ◽  
R. B. Berke

Abstract Background High-resolution Digital Image Correlation (DIC) measurements have previously been produced by stitching of neighboring images, which often requires short working distances. Separately, the image processing community has developed super resolution (SR) imaging techniques, which improve resolution by combining multiple overlapping images. Objective This work investigates the novel pairing of super resolution with digital image correlation, as an alternative method to produce high-resolution full-field strain measurements. Methods First, an image reconstruction test is performed, comparing the ability of three previously published SR algorithms to replicate a high-resolution image. Second, an applied translation is compared against DIC measurement using both low- and super-resolution images. Third, a ring sample is mechanically deformed and DIC strain measurements from low- and super-resolution images are compared. Results SR measurements show improvements compared to low-resolution images, although they do not perfectly replicate the high-resolution image. SR-DIC demonstrates reduced error and improved confidence in measuring rigid body translation when compared to low resolution alternatives, and it also shows improvement in spatial resolution for strain measurements of ring deformation. Conclusions Super resolution imaging can be effectively paired with Digital Image Correlation, offering improved spatial resolution, reduced error, and increased measurement confidence.


Author(s):  
F. Pineda ◽  
V. Ayma ◽  
C. Beltran

Abstract. High-resolution satellite images have always been in high demand due to the greater detail and precision they offer, as well as the wide scope of the fields in which they could be applied; however, satellites in operation offering very high-resolution (VHR) images has experienced an important increase, but they remain as a smaller proportion against existing lower resolution (HR) satellites. Recent models of convolutional neural networks (CNN) are very suitable for applications with image processing, like resolution enhancement of images; but in order to obtain an acceptable result, it is important, not only to define the kind of CNN architecture but the reference set of images to train the model. Our work proposes an alternative to improve the spatial resolution of HR images obtained by Sentinel-2 satellite by using the VHR images from PeruSat1, a peruvian satellite, which serve as the reference for the super-resolution approach implementation based on a Generative Adversarial Network (GAN) model, as an alternative for obtaining VHR images. The VHR PeruSat-1 image dataset is used for the training process of the network. The results obtained were analyzed considering the Peak Signal to Noise Ratios (PSNR) and the Structural Similarity (SSIM). Finally, some visual outcomes, over a given testing dataset, are presented so the performance of the model could be analyzed as well.


2019 ◽  
Vol 11 (4) ◽  
pp. 403 ◽  
Author(s):  
Weijia Li ◽  
Conghui He ◽  
Jiarui Fang ◽  
Juepeng Zheng ◽  
Haohuan Fu ◽  
...  

Automatic extraction of building footprints from high-resolution satellite imagery has become an important and challenging research issue receiving greater attention. Many recent studies have explored different deep learning-based semantic segmentation methods for improving the accuracy of building extraction. Although they record substantial land cover and land use information (e.g., buildings, roads, water, etc.), public geographic information system (GIS) map datasets have rarely been utilized to improve building extraction results in existing studies. In this research, we propose a U-Net-based semantic segmentation method for the extraction of building footprints from high-resolution multispectral satellite images using the SpaceNet building dataset provided in the DeepGlobe Satellite Challenge of IEEE Conference on Computer Vision and Pattern Recognition 2018 (CVPR 2018). We explore the potential of multiple public GIS map datasets (OpenStreetMap, Google Maps, and MapWorld) through integration with the WorldView-3 satellite datasets in four cities (Las Vegas, Paris, Shanghai, and Khartoum). Several strategies are designed and combined with the U-Net–based semantic segmentation model, including data augmentation, post-processing, and integration of the GIS map data and satellite images. The proposed method achieves a total F1-score of 0.704, which is an improvement of 1.1% to 12.5% compared with the top three solutions in the SpaceNet Building Detection Competition and 3.0% to 9.2% compared with the standard U-Net–based method. Moreover, the effect of each proposed strategy and the possible reasons for the building footprint extraction results are analyzed substantially considering the actual situation of the four cities.


2012 ◽  
Vol 170-173 ◽  
pp. 2803-2807
Author(s):  
Yan Hua Sun ◽  
Ping Wang

High resolution remote sensing images generally refer to image to the spatial resolution within 10m aerospace、aviation remote sensing images. The emergence of high-resolution images strengthened the ability to recognize the large scale features, especially for the extraction of houses information in mining area. High spatial resolution image has rich delicate texture feature, it is urgent to solution the problem of how to extract the features. The technology is very useful for statistic houses information、village relocation assessment and research of pressure coal status, providing important data basis for village relocation, statistics, assessment. Taking henan as a mining area for example, houses information extraction methods are discussed. This paper mainly research contents as followings: It is combined with the space texture information of high resolution imaging rich, using different methods to extract building information, including followings: First, ordinary image segmentation technology; this method is simple and feasible, but extracted housing information is not accurate. Second, the object-oriented method of feature extraction technology, visualization degree and extracting accuracy of this method is higher; Third, it has conducted the preliminary height extraction of the houses; according to the solar altitude angles and the shadow of the houses to calculate the height of the houses. And considering the influence of undulating terrain, using the terrain DEM data to analyze study area, finally determined the shadow length, and then used solar altitude angles to calculate houses height. Based on the verification, accuracy evaluation results show that houses contour information extraction accuracy is: accuracy of the number and area is over 80%, the total rate of wrong classifications is lower. Houses highly information extraction accuracy is within the 85%. The research methods are effective.


2011 ◽  
Vol 08 (04) ◽  
pp. 273-280
Author(s):  
YUXIANG YANG ◽  
ZENGFU WANG

This paper describes a successful application of Matting Laplacian Matrix to the problem of generating high-resolution range images. The Matting Laplacian Matrix in this paper exploits the fact that discontinuities in range and coloring tend to co-align, which enables us to generate high-resolution range image by integrating regular camera image into the range data. Using one registered and potentially high-resolution camera image as reference, we iteratively refine the input low-resolution range image, in terms of both spatial resolution and depth precision. We show that by using such a Matting Laplacian Matrix, we can get high-quality high-resolution range images.


2018 ◽  
Vol 10 (10) ◽  
pp. 1574 ◽  
Author(s):  
Dongsheng Gao ◽  
Zhentao Hu ◽  
Renzhen Ye

Due to sensor limitations, hyperspectral images (HSIs) are acquired by hyperspectral sensors with high-spectral-resolution but low-spatial-resolution. It is difficult for sensors to acquire images with high-spatial-resolution and high-spectral-resolution simultaneously. Hyperspectral image super-resolution tries to enhance the spatial resolution of HSI by software techniques. In recent years, various methods have been proposed to fuse HSI and multispectral image (MSI) from an unmixing or a spectral dictionary perspective. However, these methods extract the spectral information from each image individually, and therefore ignore the cross-correlation between the observed HSI and MSI. It is difficult to achieve high-spatial-resolution while preserving the spatial-spectral consistency between low-resolution HSI and high-resolution HSI. In this paper, a self-dictionary regression based method is proposed to utilize cross-correlation between the observed HSI and MSI. Both the observed low-resolution HSI and MSI are simultaneously considered to estimate the endmember dictionary and the abundance code. To preserve the spectral consistency, the endmember dictionary is extracted by performing a common sparse basis selection on the concatenation of observed HSI and MSI. Then, a consistent constraint is exploited to ensure the spatial consistency between the abundance code of low-resolution HSI and the abundance code of high-resolution HSI. Extensive experiments on three datasets demonstrate that the proposed method outperforms the state-of-the-art methods.


Sign in / Sign up

Export Citation Format

Share Document