A ROBUST MOVING BODY RECOGNITION METHOD

2012 ◽  
Vol 11 (02) ◽  
pp. 107-114
Author(s):  
SHOUJIA WANG ◽  
WENHUI LI ◽  
BO FU ◽  
HONGYIN NI ◽  
CONG WANG

At present, moving body recognition is one of the most active areas of research in the field of computer vision and is used widely in all kinds of videos. But the recognition accuracy of these methods has changed negatively because of the complexity of the background. In this paper, we put forward a robust recognition method. First, we obtain the moving body by tripling the temporal difference method. And then we eliminate noise from these images by mathematical morphology. Finally, we use three-scanning notation method to mark and connect the connected domain. This new method is more accurate and requires less computation in real-time experiments. The experiment result also proves its robustness.

2018 ◽  
Vol 8 (10) ◽  
pp. 1857 ◽  
Author(s):  
Jing Yang ◽  
Shaobo Li ◽  
Zong Gao ◽  
Zheng Wang ◽  
Wei Liu

The complexity of the background and the similarities between different types of precision parts, especially in the high-speed movement of conveyor belts in complex industrial scenes, pose immense challenges to the object recognition of precision parts due to diversity in illumination. This study presents a real-time object recognition method for 0.8 cm darning needles and KR22 bearing machine parts under a complex industrial background. First, we propose an image data increase algorithm based on directional flip, and we establish two types of dataset, namely, real data and increased data. We focus on increasing recognition accuracy and reducing computation time, and we design a multilayer feature fusion network to obtain feature information. Subsequently, we propose an accurate method for classifying precision parts on the basis of non-maximal suppression, and then form an improved You Only Look Once (YOLO) V3 network. We implement this method and compare it with models in our real-time industrial object detection experimental platform. Finally, experiments on real and increased datasets show that the proposed method outperforms the YOLO V3 algorithm in terms of recognition accuracy and robustness.


Author(s):  
Masud Pervej ◽  
Sabuj Das ◽  
Md. Parvez Hossain ◽  
Md. Atikuzzaman ◽  
Md. Mahin ◽  
...  

Computer vision-based recognition of Bangle vehicle license plates (LPs) is an arduous task in dirty and muddy situations. This paper proposes an efficient method for real-time computer vision-based recognition of Bangla vehicle LPs using contour analysis and prediction algorithms. The method initially applies gray scaling the input RGB images, histogram equalization to improve the grayscale image quality, edge detection using Sobel edge detector, and adaptive thresholding to convert it to a binary image. The system localizes the vehicle LP based on the maximum rectangular contour area and converts it into a predefined size. Noise removal technique using morphological dilation and erosion operation is used, followed by Gaussian filtering on binary image to improve the image quality further. The system clusters the two-lined LP into seven clusters. The sub-clustering is applied on specific clusters and makes 68 individual sub-clusters. The system extracts vector contour (VC) from each 68 individual classes. After VC extraction, the system normalizes it into a q predefined length. The system applies inter co-relation function (ICF) to categorize each sub-cluster to its previously defined individual class. For that, it calculates the maximum similarity between test and previously trained VCs. The system applies the dependency prediction algorithm in parallel to predict the whole string (district name) in the cluster-1 based on previously categorized class or classes (starting character or characters of the district part). [Formula: see text] (Metro) or (null) from cluster-2, “-” (hyphen) from cluster-3 and 6 are predicted automatically as their positions are fixed. The system is trained using 68 classes in which each class contains 100 samples generated by the augmentation technique. The system is tested using another set of 68 classes with a total of [Formula: see text] images acquiring the recognition accuracy of 96.62% with the mean computational cost of 8.363[Formula: see text]ms/f. The system is also tested using 500 vehicle whole Bangla LPs achieving the mean whole LP recognition accuracy of 95.41% with a mean computational cost of 35.803[Formula: see text]ms/f.


2021 ◽  
Vol 11 (11) ◽  
pp. 4940
Author(s):  
Jinsoo Kim ◽  
Jeongho Cho

The field of research related to video data has difficulty in extracting not only spatial but also temporal features and human action recognition (HAR) is a representative field of research that applies convolutional neural network (CNN) to video data. The performance for action recognition has improved, but owing to the complexity of the model, some still limitations to operation in real-time persist. Therefore, a lightweight CNN-based single-stream HAR model that can operate in real-time is proposed. The proposed model extracts spatial feature maps by applying CNN to the images that develop the video and uses the frame change rate of sequential images as time information. Spatial feature maps are weighted-averaged by frame change, transformed into spatiotemporal features, and input into multilayer perceptrons, which have a relatively lower complexity than other HAR models; thus, our method has high utility in a single embedded system connected to CCTV. The results of evaluating action recognition accuracy and data processing speed through challenging action recognition benchmark UCF-101 showed higher action recognition accuracy than the HAR model using long short-term memory with a small amount of video frames and confirmed the real-time operational possibility through fast data processing speed. In addition, the performance of the proposed weighted mean-based HAR model was verified by testing it in Jetson NANO to confirm the possibility of using it in low-cost GPU-based embedded systems.


Author(s):  
Phakawat Pattarapongsin ◽  
Bipul Neupane ◽  
Jirayus Vorawan ◽  
Harit Sutthikulsombat ◽  
Teerayut Horanont

Sensors ◽  
2021 ◽  
Vol 21 (5) ◽  
pp. 1919
Author(s):  
Shuhua Liu ◽  
Huixin Xu ◽  
Qi Li ◽  
Fei Zhang ◽  
Kun Hou

With the aim to solve issues of robot object recognition in complex scenes, this paper proposes an object recognition method based on scene text reading. The proposed method simulates human-like behavior and accurately identifies objects with texts through careful reading. First, deep learning models with high accuracy are adopted to detect and recognize text in multi-view. Second, datasets including 102,000 Chinese and English scene text images and their inverse are generated. The F-measure of text detection is improved by 0.4% and the recognition accuracy is improved by 1.26% because the model is trained by these two datasets. Finally, a robot object recognition method is proposed based on the scene text reading. The robot detects and recognizes texts in the image and then stores the recognition results in a text file. When the user gives the robot a fetching instruction, the robot searches for corresponding keywords from the text files and achieves the confidence of multiple objects in the scene image. Then, the object with the maximum confidence is selected as the target. The results show that the robot can accurately distinguish objects with arbitrary shape and category, and it can effectively solve the problem of object recognition in home environments.


Sign in / Sign up

Export Citation Format

Share Document