Multi-scale personnel deep feature detection algorithm based on Extended-YOLOv3

2021 ◽  
Vol 40 (1) ◽  
pp. 773-786
Author(s):  
Shuai Liu ◽  
Ying Xu ◽  
Lingming Guo ◽  
Meng Shao ◽  
Guodong Yue ◽  
...  

Tens of thousands of work-related injuries and deaths are reported in the construction industry each year, and a high percentage of them are due to construction workers not wearing safety equipment. In order to address this safety issue, it is particularly necessary to automatically identify people and detect the safety characteristics of personnel at the same time in the prefabricated building. Therefore, this paper proposes a depth feature detection algorithm based on the Extended-YOLOv3 model. On the basis of the YOLOv3 network, a security feature recognition network and a feature transmission network are added to achieve the purpose of detecting security features while identifying personnel. Firstly, a security feature recognition network is added side by side on the basis of the YOLOv3 network to analyze the wearing characteristics of construction workers. Secondly, the S-SPP module is added to the object detection and feature recognition network to broaden the features of the deep network and help the network extract more useful features from the high-resolution input image. Finally, a special feature transmission network is designed to transfer features between the construction worker detection network and the security feature recognition network, so that the two networks can obtain feature information from the other network respectively. Compared with YOLOv3 algorithm, Extended-YOLOv3 in this paper adds security feature recognition and feature transmission functions, and adds S-SPP module to the object detection and feature recognition network. The experimental results show that the Extended-YOLOv3 algorithm is 1.3% better than the YOLOV3 algorithm in AP index.

Author(s):  
A. S. Tuzhilkin

The paper deals with the feature detection problem in aircraft machine vision systems. We developed a fast detection algorithm. We determined feature locations in an image sequence depicting fast-moving objects as registered by aircraft machine vision systems. We analysed existing feature detection algorithms. The input image frame sequence undergoes a geometric transformation, that is, rotation. We simulated how our algorithm processed a sequence of frames depicting the target. We compared efficiency and speed of feature detection in the well-known SIFT algorithm with that of our algorithm. We demonstrate that the algorithm developed ensures faster feature detection that is more resistant to geometric transformations of the target image as compared to the SIFT algorithm.


2021 ◽  
Vol 252 ◽  
pp. 01048
Author(s):  
Zhishan Zhu ◽  
Aike Liu ◽  
Haifei Chi

This paper designs a simple non-contact temperature measurement and identity recognition device based on multiple embedded control systems and feature recognition algorithms. The device can achieve multiple functions such as non-contact temperature measurement, face recognition, mask recognition, and smart alarm. The system consists of three parts: main control, interactive system and detection system: the main control selects STM32F407VGT6 to process the data returned by multiple sensors and realize the mutual communication of each system; the interactive system uses the HMI serial touch screen to realize the visualization of data and human Machine operation function; the detection system is equipped with MLX90614 temperature detection module and OpenMV4 machine vision module to realize functions such as temperature detection, face recognition and mask recognition[1]. In addition, in order to ensure the accuracy and stability of the detection results, this article specifically designs temperature data filtering and compensation algorithms, LBP feature detection algorithms and other intelligent algorithms. Through experiments, the accuracy of this device to detect 28°C-48°C is within 0.8°C, and the accuracy of identifying faces and masks is above 98%.


2017 ◽  
Vol 2 (1) ◽  
pp. 80-87
Author(s):  
Puyda V. ◽  
◽  
Stoian. A.

Detecting objects in a video stream is a typical problem in modern computer vision systems that are used in multiple areas. Object detection can be done on both static images and on frames of a video stream. Essentially, object detection means finding color and intensity non-uniformities which can be treated as physical objects. Beside that, the operations of finding coordinates, size and other characteristics of these non-uniformities that can be used to solve other computer vision related problems like object identification can be executed. In this paper, we study three algorithms which can be used to detect objects of different nature and are based on different approaches: detection of color non-uniformities, frame difference and feature detection. As the input data, we use a video stream which is obtained from a video camera or from an mp4 video file. Simulations and testing of the algoritms were done on a universal computer based on an open-source hardware, built on the Broadcom BCM2711, quad-core Cortex-A72 (ARM v8) 64-bit SoC processor with frequency 1,5GHz. The software was created in Visual Studio 2019 using OpenCV 4 on Windows 10 and on a universal computer operated under Linux (Raspbian Buster OS) for an open-source hardware. In the paper, the methods under consideration are compared. The results of the paper can be used in research and development of modern computer vision systems used for different purposes. Keywords: object detection, feature points, keypoints, ORB detector, computer vision, motion detection, HSV model color


Author(s):  
Samuel Humphries ◽  
Trevor Parker ◽  
Bryan Jonas ◽  
Bryan Adams ◽  
Nicholas J Clark

Quick identification of building and roads is critical for execution of tactical US military operations in an urban environment. To this end, a gridded, referenced, satellite images of an objective, often referred to as a gridded reference graphic or GRG, has become a standard product developed during intelligence preparation of the environment. At present, operational units identify key infrastructure by hand through the work of individual intelligence officers. Recent advances in Convolutional Neural Networks, however, allows for this process to be streamlined through the use of object detection algorithms. In this paper, we describe an object detection algorithm designed to quickly identify and label both buildings and road intersections present in an image. Our work leverages both the U-Net architecture as well the SpaceNet data corpus to produce an algorithm that accurately identifies a large breadth of buildings and different types of roads. In addition to predicting buildings and roads, our model numerically labels each building by means of a contour finding algorithm. Most importantly, the dual U-Net model is capable of predicting buildings and roads on a diverse set of test images and using these predictions to produce clean GRGs.


Plant Methods ◽  
2021 ◽  
Vol 17 (1) ◽  
Author(s):  
Hiranya Jayakody ◽  
Paul Petrie ◽  
Hugo Jan de Boer ◽  
Mark Whitty

Abstract Background Stomata analysis using microscope imagery provides important insight into plant physiology, health and the surrounding environmental conditions. Plant scientists are now able to conduct automated high-throughput analysis of stomata in microscope data, however, existing detection methods are sensitive to the appearance of stomata in the training images, thereby limiting general applicability. In addition, existing methods only generate bounding-boxes around detected stomata, which require users to implement additional image processing steps to study stomata morphology. In this paper, we develop a fully automated, robust stomata detection algorithm which can also identify individual stomata boundaries regardless of the plant species, sample collection method, imaging technique and magnification level. Results The proposed solution consists of three stages. First, the input image is pre-processed to remove any colour space biases occurring from different sample collection and imaging techniques. Then, a Mask R-CNN is applied to estimate individual stomata boundaries. The feature pyramid network embedded in the Mask R-CNN is utilised to identify stomata at different scales. Finally, a statistical filter is implemented at the Mask R-CNN output to reduce the number of false positive generated by the network. The algorithm was tested using 16 datasets from 12 sources, containing over 60,000 stomata. For the first time in this domain, the proposed solution was tested against 7 microscope datasets never seen by the algorithm to show the generalisability of the solution. Results indicated that the proposed approach can detect stomata with a precision, recall, and F-score of 95.10%, 83.34%, and 88.61%, respectively. A separate test conducted by comparing estimated stomata boundary values with manually measured data showed that the proposed method has an IoU score of 0.70; a 7% improvement over the bounding-box approach. Conclusions The proposed method shows robust performance across multiple microscope image datasets of different quality and scale. This generalised stomata detection algorithm allows plant scientists to conduct stomata analysis whilst eliminating the need to re-label and re-train for each new dataset. The open-source code shared with this project can be directly deployed in Google Colab or any other Tensorflow environment.


Author(s):  
Louis Lecrosnier ◽  
Redouane Khemmar ◽  
Nicolas Ragot ◽  
Benoit Decoux ◽  
Romain Rossi ◽  
...  

This paper deals with the development of an Advanced Driver Assistance System (ADAS) for a smart electric wheelchair in order to improve the autonomy of disabled people. Our use case, built from a formal clinical study, is based on the detection, depth estimation, localization and tracking of objects in wheelchair’s indoor environment, namely: door and door handles. The aim of this work is to provide a perception layer to the wheelchair, enabling this way the detection of these keypoints in its immediate surrounding, and constructing of a short lifespan semantic map. Firstly, we present an adaptation of the YOLOv3 object detection algorithm to our use case. Then, we present our depth estimation approach using an Intel RealSense camera. Finally, as a third and last step of our approach, we present our 3D object tracking approach based on the SORT algorithm. In order to validate all the developments, we have carried out different experiments in a controlled indoor environment. Detection, distance estimation and object tracking are experimented using our own dataset, which includes doors and door handles.


Sensors ◽  
2021 ◽  
Vol 21 (6) ◽  
pp. 2052
Author(s):  
Xinghai Yang ◽  
Fengjiao Wang ◽  
Zhiquan Bai ◽  
Feifei Xun ◽  
Yulin Zhang ◽  
...  

In this paper, a deep learning-based traffic state discrimination method is proposed to detect traffic congestion at urban intersections. The detection algorithm includes two parts, global speed detection and a traffic state discrimination algorithm. Firstly, the region of interest (ROI) is selected as the road intersection from the input image of the You Only Look Once (YOLO) v3 object detection algorithm for vehicle target detection. The Lucas-Kanade (LK) optical flow method is employed to calculate the vehicle speed. Then, the corresponding intersection state can be obtained based on the vehicle speed and the discrimination algorithm. The detection of the vehicle takes the position information obtained by YOLOv3 as the input of the LK optical flow algorithm and forms an optical flow vector to complete the vehicle speed detection. Experimental results show that the detection algorithm can detect the vehicle speed and traffic state discrimination method can judge the traffic state accurately, which has a strong anti-interference ability and meets the practical application requirements.


Sign in / Sign up

Export Citation Format

Share Document