scene segmentation
Recently Published Documents


TOTAL DOCUMENTS

360
(FIVE YEARS 79)

H-INDEX

29
(FIVE YEARS 7)

2022 ◽  
Vol 59 (2) ◽  
pp. 102840
Author(s):  
Xianfeng Ou ◽  
Hanpu Wang ◽  
Wujing Li ◽  
Guoyun Zhang ◽  
Siyuan Chen

2021 ◽  
Vol 2021 ◽  
pp. 1-16
Author(s):  
Yangmei Zhang

This paper is aimed at studying underwater object detection and positioning. Objects are detected and positioned through an underwater scene segmentation-based weak object detection algorithm and underwater positioning technology based on the three-dimensional (3D) omnidirectional magnetic induction smart sensor. The proposed weak object detection involves a predesigned U-shaped network- (U-Net-) architectured image segmentation network, which has been improved before application. The key factor of underwater positioning technology based on 3D omnidirectional magnetic induction is the magnetic induction intensity. The results show that the image-enhanced object detection method improves the accuracy of Yellow Croaker, Goldfish, and Mandarin Fish by 3.2%, 1.5%, and 1.6%, respectively. In terms of sensor positioning technology, under the positioning Signal-to-Noise Ratio (SNR) of 15 dB and 20 dB, the curve trends of actual distance and positioning distance are consistent, while SNR = 10   dB , the two curves deviate greatly. The research conclusions read as follows: an underwater scene segmentation-based weak object detection method is proposed for invalid underwater object samples from poor labeling, which can effectively segment the background from underwater objects, remove the negative impact of invalid samples, and improve the precision of weak object detection. The positioning model based on a 3D coil magnetic induction sensor can obtain more accurate positioning coordinates. The effectiveness of 3D omnidirectional magnetic induction coil underwater positioning technology is verified by simulation experiments.


2021 ◽  
pp. 1-10
Author(s):  
Rui Cao ◽  
Feng Jiang ◽  
Zhao Wu ◽  
Jia Ren

With the advancement of computer performance, deep learning is playing a vital role on hardware platforms. Indoor scene segmentation is a challenging deep learning task because indoor objects tend to obscure each other, and the dense layout increases the difficulty of segmentation. Still, current networks pursue accuracy improvement, sacrifice speed, and augment memory resource usage. To solve this problem, achieve a compromise between accuracy, speed, and model size. This paper proposes Multichannel Fusion Network (MFNet) for indoor scene segmentation, which mainly consists of Dense Residual Module(DRM) and Multi-scale Feature Extraction Module(MFEM). MFEM uses depthwise separable convolution to cut the number of parameters, matches different sizes of convolution kernels and dilation rates to achieve optimal receptive field; DRM fuses feature maps at several levels of resolution to optimize segmentation details. Experimental results on the NYU V2 dataset show that the proposed method achieves very competitive results compared with other advanced algorithms, with a segmentation speed of 38.47 fps, nearly twice that of Deeplab v3+, but only 1/5 of the number of parameters of Deeplab v3 + . Its segmentation results were close to those of advanced segmentation networks, making it beneficial for the real-time processing of images.


Sensors ◽  
2021 ◽  
Vol 21 (22) ◽  
pp. 7730
Author(s):  
◽  

Semantic segmentation is one of the most active research topics in computer vision with the goal to assign dense semantic labels for all pixels in a given image. In this paper, we introduce HFEN (Hierarchical Feature Extraction Network), a lightweight network to reach a balance between inference speed and segmentation accuracy. Our architecture is based on an encoder-decoder framework. The input images are down-sampled through an efficient encoder to extract multi-layer features. Then the extracted features are fused via a decoder, where the global contextual information and spatial information are aggregated for final segmentations with real-time performance. Extensive experiments have been conducted on two standard benchmarks, Cityscapes and Camvid, where our network achieved superior performance on NVIDIA 2080Ti.


Electronics ◽  
2021 ◽  
Vol 10 (15) ◽  
pp. 1851
Author(s):  
An Wang ◽  
Xiaohong Cao ◽  
Lei Lu ◽  
Xinjing Zhou ◽  
Xuecheng Sun

Human head statistics is widely used in the construction of smart cities and has great market value. In order to solve the problem of missing pedestrian features and poor statistics results in a large-angle overlooking scene, in this paper we propose a human head statistics system that consists of head detection, head tracking and head counting, where the proposed You-Only-Look-Once-Head (YOLOv5-H) network, improved from YOLOv5, is taken as the head detection benchmark, the DeepSORT algorithm with the Fusion-Hash algorithm for feature extraction (DeepSORT-FH) is proposed to track heads, and heads are counted by the proposed cross-boundary counting algorithm based on scene segmentation. Specifically, Complete-Intersection-over-Union (CIoU) is taken as the loss function of YOLOv5-H to make the predicted boxes more in line with the real boxes. The results demonstrate that the recall rate and [email protected] of the proposed YOLOv5-H can reach up to 94.3% and 93.1%, respectively, on the SCUT_HEAD dataset. The statistics system has an extremely low error rate of 3.5% on the TownCentreXVID dataset while maintaining a frame rate of 18FPS, which can meet the needs of human head statistics in monitoring scenarios and has a good application prospect.


2021 ◽  
Vol 21 (7) ◽  
pp. 13
Author(s):  
Ross Goutcher ◽  
Christian Barrington ◽  
Paul B. Hibbard ◽  
Bruce Graham

2021 ◽  
Author(s):  
Li Ding ◽  
Rini Sherony ◽  
Bruce Mehler ◽  
Bryan Reimer

Sign in / Sign up

Export Citation Format

Share Document