Lightweight deep network-enabled real-time low-visibility enhancement for promoting vessel detection in maritime video surveillance

2021 ◽  
pp. 1-21
Author(s):  
Yu Guo ◽  
Yuxu Lu ◽  
Ryan Wen Liu

Abstract Maritime video surveillance has become an essential part of the vessel traffic services system, intended to guarantee vessel traffic safety and security in maritime applications. To make maritime surveillance more feasible and practicable, many intelligent vision-empowered technologies have been developed to automatically detect moving vessels from maritime visual sensing data (i.e., maritime surveillance videos). However, when visual data is collected in a low-visibility environment, the essential optical information is often hidden in the dark, potentially resulting in decreased accuracy of vessel detection. To guarantee reliable vessel detection under low-visibility conditions, the paper proposes a low-visibility enhancement network (termed LVENet) based on Retinex theory to enhance imaging quality in maritime video surveillance. LVENet is a lightweight deep neural network incorporating a depthwise separable convolution. The synthetically-degraded image generation and hybrid loss function are further presented to enhance the robustness and generalisation capacities of LVENet. Both full-reference and no-reference evaluation experiments demonstrate that LVENet could yield comparable or even better visual qualities than other state-of-the-art methods. In addition, it takes LVENet just 0⋅0045 s to restore degraded images with size 1920 × 1080 pixels on an NVIDIA 2080Ti GPU, which can adequately meet real-time requirements. Using LVENet, vessel detection performance can be greatly improved with enhanced visibility under low-light imaging conditions.

2013 ◽  
Vol 756-759 ◽  
pp. 439-442
Author(s):  
Shang Fu Gong ◽  
Li Gang Wu ◽  
Yan Jun Wang

Video surveillance, convenience and rich information, has been widely used in security, protection, monitoring and other occasions, and has already been one of the most important precautionary measures in commercial, residential and transportation areas. However, considering the massive data transmission needs and higher real-time requirements for video surveillance, a remote video surveillance plan has been put forward on the analysis basis of Microsoft DirectShow and Image Compression. This plan adopts the C/S structure, adapts to the requirements of real-time video transmission, with a better fluency. In addition, the picture clarity satisfies the application requirements.


2013 ◽  
Vol 2013 ◽  
pp. 1-5 ◽  
Author(s):  
Sanjay Singh ◽  
A. S. Mandal ◽  
Chandra Shekhar ◽  
Anil Vohra

Change detection is one of the several important problems in the design of any automated video surveillance system. Appropriate selection of frames of significant changes can minimize the communication and processing overheads for such systems. This research presents the design of a VLSI architecture for change detection in a video sequence and its implementation on Virtex-IIPro FPGA platform. Clustering-based scheme is used for change detection. The proposed system is designed to meet the real-time requirements of video surveillance applications. It robustly detects the changes in a video stream in real time at 25 frames per second (fps) in gray scale CIF size video.


Author(s):  
Jun-hua Chen ◽  
Da-hu Wang ◽  
Cun-yuan Sun

Objective: This study focused on the application of wearable technology in the safety monitoring and early warning for subway construction workers. Methods: With the help of real-time video surveillance and RFID positioning which was applied in the construction has realized the real-time monitoring and early warning of on-site construction to a certain extent, but there are still some problems. Real-time video surveillance technology relies on monitoring equipment, while the location of the equipment is fixed, so it is difficult to meet the full coverage of the construction site. However, wearable technologies can solve this problem, they have outstanding performance in collecting workers’ information, especially physiological state data and positioning data. Meanwhile, wearable technology has no impact on work and is not subject to the inference of dynamic environment. Results and conclusion: The first time the system applied to subway construction was a great success. During the construction of the station, the number of occurrences of safety warnings was 43 times, but the number of occurrences of safety accidents was 0, which showed that the safety monitoring and early warning system played a significant role and worked out perfectly.


2021 ◽  
Vol 1 (1) ◽  
Author(s):  
E. Bertino ◽  
M. R. Jahanshahi ◽  
A. Singla ◽  
R.-T. Wu

AbstractThis paper addresses the problem of efficient and effective data collection and analytics for applications such as civil infrastructure monitoring and emergency management. Such problem requires the development of techniques by which data acquisition devices, such as IoT devices, can: (a) perform local analysis of collected data; and (b) based on the results of such analysis, autonomously decide further data acquisition. The ability to perform local analysis is critical in order to reduce the transmission costs and latency as the results of an analysis are usually smaller in size than the original data. As an example, in case of strict real-time requirements, the analysis results can be transmitted in real-time, whereas the actual collected data can be uploaded later on. The ability to autonomously decide about further data acquisition enhances scalability and reduces the need of real-time human involvement in data acquisition processes, especially in contexts with critical real-time requirements. The paper focuses on deep neural networks and discusses techniques for supporting transfer learning and pruning, so to reduce the times for training the networks and the size of the networks for deployment at IoT devices. We also discuss approaches based on machine learning reinforcement techniques enhancing the autonomy of IoT devices.


Author(s):  
Yuefeng Wang ◽  
Kuang Mao ◽  
Tong Chen ◽  
Yanglong Yin ◽  
Shuibing He ◽  
...  

2021 ◽  
Vol 20 (3) ◽  
pp. 1-22
Author(s):  
David Langerman ◽  
Alan George

High-resolution, low-latency apps in computer vision are ubiquitous in today’s world of mixed-reality devices. These innovations provide a platform that can leverage the improving technology of depth sensors and embedded accelerators to enable higher-resolution, lower-latency processing for 3D scenes using depth-upsampling algorithms. This research demonstrates that filter-based upsampling algorithms are feasible for mixed-reality apps using low-power hardware accelerators. The authors parallelized and evaluated a depth-upsampling algorithm on two different devices: a reconfigurable-logic FPGA embedded within a low-power SoC; and a fixed-logic embedded graphics processing unit. We demonstrate that both accelerators can meet the real-time requirements of 11 ms latency for mixed-reality apps. 1


2021 ◽  
Vol 413 ◽  
pp. 125291
Author(s):  
Min Liu ◽  
Haoran Li ◽  
Lan Bai ◽  
Kai Zheng ◽  
Zhipeng Zhao ◽  
...  

2020 ◽  
Vol 13 (1) ◽  
pp. 89
Author(s):  
Manuel Carranza-García ◽  
Jesús Torres-Mateo ◽  
Pedro Lara-Benítez ◽  
Jorge García-Gutiérrez

Object detection using remote sensing data is a key task of the perception systems of self-driving vehicles. While many generic deep learning architectures have been proposed for this problem, there is little guidance on their suitability when using them in a particular scenario such as autonomous driving. In this work, we aim to assess the performance of existing 2D detection systems on a multi-class problem (vehicles, pedestrians, and cyclists) with images obtained from the on-board camera sensors of a car. We evaluate several one-stage (RetinaNet, FCOS, and YOLOv3) and two-stage (Faster R-CNN) deep learning meta-architectures under different image resolutions and feature extractors (ResNet, ResNeXt, Res2Net, DarkNet, and MobileNet). These models are trained using transfer learning and compared in terms of both precision and efficiency, with special attention to the real-time requirements of this context. For the experimental study, we use the Waymo Open Dataset, which is the largest existing benchmark. Despite the rising popularity of one-stage detectors, our findings show that two-stage detectors still provide the most robust performance. Faster R-CNN models outperform one-stage detectors in accuracy, being also more reliable in the detection of minority classes. Faster R-CNN Res2Net-101 achieves the best speed/accuracy tradeoff but needs lower resolution images to reach real-time speed. Furthermore, the anchor-free FCOS detector is a slightly faster alternative to RetinaNet, with similar precision and lower memory usage.


Sign in / Sign up

Export Citation Format

Share Document