Towards Real-time CNN Inference from a Video Stream on a Mobile GPU (WiP Paper)

Author(s):  
Chanyoung Oh ◽  
Gunju Park ◽  
Sumin Kim ◽  
Dohee Kim ◽  
Youngmin Yi
Keyword(s):  
1999 ◽  
pp. 71-84 ◽  
Author(s):  
G. Medioni ◽  
G. Guy ◽  
H. Rom ◽  
A. François
Keyword(s):  

2020 ◽  
Vol 21 (3) ◽  
pp. 181-190
Author(s):  
Jaroslav Frnda ◽  
Marek Durica ◽  
Mihail Savrasovs ◽  
Philippe Fournier-Viger ◽  
Jerry Chun-Wei Lin

AbstractThis paper deals with an analysis of Kohonen map usage possibility for real-time evaluation of end-user video quality perception. The Quality of Service framework (QoS) describes how the network impairments (network utilization or packet loss) influence the picture quality, but it does not reflect precisely on customer subjective perceived quality of received video stream. There are several objective video assessment metrics based on mathematical models trying to simulate human visual system but each of them has its own evaluation scale. This causes a serious problem for service providers to identify a critical point when intervention into the network behaviour is needed. On the other hand, subjective tests (Quality of Experience concept) are time-consuming and costly and of course, cannot be performed in real-time. Therefore, we proposed a mapping function able to predict subjective end-user quality perception based on the situation in a network, video stream features and results obtained from the objective video assessment method.


2020 ◽  
Author(s):  
Krzysztof Blachut ◽  
Hubert Szolc ◽  
Mateusz Wasala ◽  
Tomasz Kryjak ◽  
Marek Gorgon

In this paper we present a vision based hardware-software control system enabling autonomous landing of a mul-tirotor unmanned aerial vehicle (UAV). It allows the detection of a marked landing pad in real-time for a 1280 x 720 @ 60 fps video stream. In addition, a LiDAR sensor is used to measure the altitude above ground. A heterogeneous Zynq SoC device is used as the computing platform. The solution was tested on a number of sequences and the landing pad was detected with 96% accuracy. This research shows that a reprogrammable heterogeneous computing system is a good solution for UAVs because it enables real-time data stream processing with relatively low energy consumption.


2021 ◽  
Author(s):  
Daniele Berardini ◽  
Adriano Mancini ◽  
Primo Zingaretti ◽  
Sara Moccia

Abstract Nowadays, video surveillance has a crucial role. Analyzing surveillance videos is, however, a time consuming and tiresome procedure. In the last years, artificial intelligence paved the way for automatic and accurate surveillance-video analysis. In parallel to the development of artificial-intelligence methodologies, edge computing is becoming an active field of research with the final goal to provide cost-effective and real time deployment of the developed methodologies. In this work, we present an edge artificial intelligence application to video surveillance. Our approach relies on a set of four IP cameras, which acquire video frames that are processed on the edge using the NVIDIA® Jetson Nano. A state-of-the-art deep-learning model, called Single Shot multibox Detector (SSD) MobileNetV2 network, is used to perform object and people detection in real-time. The proposed infrastructure obtained an inference speed of ∼10.0 Frames per Second (FPS) for each parallel video stream. These results prompt the possibility of translating our work into a real word scenario. The integration of the presented application into a wider monitoring system with a central unit could bring benefits to the overall infrastructure. Indeed our application could send only video-related high-level information to the central unit, allowing it to combine information with data coming from other sensing devices without unuseful data overload. This would ensure a fast response in case of emergency or detected anomalies. We hope this work will contribute to stimulate the research in the field of edge artificial intelligence for video surveillance.


Biometrics ◽  
2017 ◽  
pp. 761-777
Author(s):  
Di Zhao

Mobile GPU computing, or System on Chip with embedded GPU (SoC GPU), becomes in great demand recently. Since these SoCs are designed for mobile devices with real-time applications such as image processing and video processing, high-efficient implementations of wavelet transform are essential for these chips. In this paper, the author develops two SoC GPU based DWT: signal based parallelization for discrete wavelet transform (sDWT) and coefficient based parallelization for discrete wavelet transform (cDWT), and the author evaluates the performance of three-dimensional wavelet transform on SoC GPU Tegra K1. Computational results show that, SoC GPU based DWT is significantly faster than SoC CPU based DWT. Computational results also show that, sDWT can generally satisfy the requirement of real-time processing (30 frames per second) with the image sizes of 352×288, 480×320, 720×480 and 1280×720, while cDWT can only obtain read-time processing with small image sizes of 352×288 and 480×320.


Author(s):  
Sanjay Singh ◽  
Sumeet Saurav ◽  
Ravi Saini ◽  
Anil K Saini ◽  
Chandra Shekhar ◽  
...  
Keyword(s):  

Sign in / Sign up

Export Citation Format

Share Document