scholarly journals Multi Receptive Field Network for Semantic Segmentation

Author(s):  
Jianlong Yuan ◽  
Zelu Deng ◽  
Shu Wang ◽  
Zhenbo Luo
2021 ◽  
Vol 13 (23) ◽  
pp. 4902
Author(s):  
Guanzhou Chen ◽  
Xiaoliang Tan ◽  
Beibei Guo ◽  
Kun Zhu ◽  
Puyun Liao ◽  
...  

Semantic segmentation is a fundamental task in remote sensing image analysis (RSIA). Fully convolutional networks (FCNs) have achieved state-of-the-art performance in the task of semantic segmentation of natural scene images. However, due to distinctive differences between natural scene images and remotely-sensed (RS) images, FCN-based semantic segmentation methods from the field of computer vision cannot achieve promising performances on RS images without modifications. In previous work, we proposed an RS image semantic segmentation framework SDFCNv1, combined with a majority voting postprocessing method. Nevertheless, it still has some drawbacks, such as small receptive field and large number of parameters. In this paper, we propose an improved semantic segmentation framework SDFCNv2 based on SDFCNv1, to conduct optimal semantic segmentation on RS images. We first construct a novel FCN model with hybrid basic convolutional (HBC) blocks and spatial-channel-fusion squeeze-and-excitation (SCFSE) modules, which occupies a larger receptive field and fewer network model parameters. We also put forward a data augmentation method based on spectral-specific stochastic-gamma-transform-based (SSSGT-based) during the model training process to improve generalizability of our model. Besides, we design a mask-weighted voting decision fusion postprocessing algorithm for image segmentation on overlarge RS images. We conducted several comparative experiments on two public datasets and a real surveying and mapping dataset. Extensive experimental results demonstrate that compared with the SDFCNv1 framework, our SDFCNv2 framework can increase the mIoU metric by up to 5.22% while only using about half of parameters.


2022 ◽  
Vol 2161 (1) ◽  
pp. 012064
Author(s):  
M Dhruv ◽  
R Sai Chandra Teja ◽  
R Sri Devi ◽  
S Nagesh Kumar

Abstract COVID-19 is an emerging infectious disease that has been rampant worldwide since its onset causing Lung irregularity and severe respiratory failure due to pneumonia. The Community-Acquired Pneumonia (CAP), Normal, and COVID-19 Computed Tomography (CT) scan images are classified using Involution Receptive Field Network from Large COVID-19 CT scan slice dataset. The proposed lightweight Involution Receptive Field Network (InRFNet) is spatial specific and channel-agnostic with Receptive Field structure to enhance the feature map extraction. The InRFNet model evaluation results show high training (99%) and validation (96%) accuracy. The performance metrics of the InRFNet model are Sensitivity (94.48%), Specificity (97.87%), Recall (96.34%), F1-score (96.33%), kappa score (94.10%), ROC-AUC (99.41%), mean square error (0.04), and the total number of parameters (33100).


2020 ◽  
Vol 24 (11) ◽  
pp. 3215-3225 ◽  
Author(s):  
Liangliang Liu ◽  
Fang-Xiang Wu ◽  
Yu-Ping Wang ◽  
Jianxin Wang

2021 ◽  
Vol 2003 (1) ◽  
pp. 012006
Author(s):  
Yu Dong ◽  
Hongbo Wang ◽  
Jingjing Luo ◽  
Zhiping Lai ◽  
Fuhao Wang ◽  
...  

Author(s):  
Xin Tong ◽  
Xianghua Ying ◽  
Yongjie Shi ◽  
He Zhao ◽  
Ruibin Wang

Several images are taken for the same scene with many view directions. Given a pixel in any one image of them, its correspondences may appear in the other images. However, by using existing semantic segmentation methods, we find that the pixel and its correspondences do not always have the same inferred label as expected. Fortunately, from the knowledge of multiple view geometry, if we keep the position of a camera unchanged, and only vary its orientation, there is a homography transformation to describe the relationship of corresponding pixels in such images. Based on this fact, we propose to generate images which are the same as real images of the scene taken in certain novel view directions for training and evaluation. We also introduce gradient guided deformable convolution to alleviate the inconsistency, by learning dynamic proper receptive field from feature gradients. Furthermore, a novel consistency loss is presented to enforce feature consistency. Compared with previous approaches, the proposed method gets significant improvement in both cross-view consistency and semantic segmentation performance on images with abundant view directions, while keeping comparable or better performance on the existing datasets.


Sign in / Sign up

Export Citation Format

Share Document