Automatic Badminton Action Recognition Using RGB-D Sensor

2014 ◽  
Vol 1042 ◽  
pp. 89-93 ◽  
Author(s):  
H.Y. Ting ◽  
K.S. Sim ◽  
F.S. Abas

This paper presents a method to recognize badminton action from depth map sequences acquired by Microsoft Kinect sensor. Badminton is one of Malaysia’s most popular, but there is still lack of research on action recognition focusing on this sport. In this research, bone orientation details of badminton players are computed and extracted in order to form a bag of quaternions feature vectors. After conversion to log-covariance matrix, the system is trained and the badminton actions are classified by a support vector machine classifier. Our experimental dataset of depth map sequences composed of 300 badminton action samples of 10 badminton actions performed by six badminton players. The dataset varies in terms of human body size, clothes, speed, and gender. Experimental result has shown that nearly 92% of average recognition accuracy (ARA) was achieved in inter-class leave one sample out cross validation test. At the same time, 86% of ARA was achieved in inter-class cross subject validation test.

2020 ◽  
Vol 5 (2) ◽  
pp. 504
Author(s):  
Matthias Omotayo Oladele ◽  
Temilola Morufat Adepoju ◽  
Olaide ` Abiodun Olatoke ◽  
Oluwaseun Adewale Ojo

Yorùbá language is one of the three main languages that is been spoken in Nigeria. It is a tonal language that carries an accent on the vowel alphabets. There are twenty-five (25) alphabets in Yorùbá language with one of the alphabets a digraph (GB). Due to the difficulty in typing handwritten Yorùbá documents, there is a need to develop a handwritten recognition system that can convert the handwritten texts to digital format. This study discusses the offline Yorùbá handwritten word recognition system (OYHWR) that recognizes Yorùbá uppercase alphabets. Handwritten characters and words were obtained from different writers using the paint application and M708 graphics tablets. The characters were used for training and the words were used for testing. Pre-processing was done on the images and the geometric features of the images were extracted using zoning and gradient-based feature extraction. Geometric features are the different line types that form a particular character such as the vertical, horizontal, and diagonal lines. The geometric features used are the number of horizontal lines, number of vertical lines, number of right diagonal lines, number of left diagonal lines, total length of all horizontal lines, total length of all vertical lines, total length of all right slanting lines, total length of all left-slanting lines and the area of the skeleton. The characters are divided into 9 zones and gradient feature extraction was used to extract the horizontal and vertical components and geometric features in each zone. The words were fed into the support vector machine classifier and the performance was evaluated based on recognition accuracy. Support vector machine is a two-class classifier, hence a multiclass SVM classifier least square support vector machine (LSSVM) was used for word recognition. The one vs one strategy and RBF kernel were used and the recognition accuracy obtained from the tested words ranges between 66.7%, 83.3%, 85.7%, 87.5%, and 100%. The low recognition rate for some of the words could be as a result of the similarity in the extracted features.


security access control systems and forensic applications. Performance of conventional unimodal biometric systems is generally suffered due to the noisy data, non universality and intolerable error rate. In propose system, multi layer Convolutional Neural Network (CNN) is applied to multimodal biometric human authentication using face, palm vein and fingerprints to increase the robustness of system. For the classification linear Support Vector Machine classifier is used. For the evaluation of system self developed face, palm vein and fingerprint database having 4,500 images are used. The performance of the system is evaluated on the basis of % recognition accuracy, and it shows significant improvement over the unimodal-biometric system and existing multimodal systems.


2011 ◽  
Vol 121-126 ◽  
pp. 1952-1956
Author(s):  
Rui Hu Wang

The automatic classification of erythrocyte is critical to clinic blood-related disease treatment in Medical Image Computer Aided Diagnosing(MICAD). After 3D height field recovered from the varied shading, the depth map of each point on the surfaces is applied to calculate Gaussian curvature and mean curvature, which are used to produce surface type label image. Accordingly the surface is segmented into different parts through multi-scale bivariate polynomials function fitting. The count of different surface types is used to design a classifier for training and classifing the red blood cell by means of support vector machine and particle swarm optimization. The experimental result shows that this approach is easily to implement and promising.


2014 ◽  
Vol 609-610 ◽  
pp. 1448-1452
Author(s):  
Kun Zhang ◽  
Min Rui Fei

Features analysis is an important task which can significantly affect the performance of automatic bacteria colony picking. This paper presents a novel approach for adaptive colony segmentation by classifying the detected peaks of intensity histograms of images. The relevance and importance of these features can be determined in an improved support vector machine classifier using unascertained least square estimation. Experimental results show that the proposed unascertained support vector machine (USVM) has better recognition accuracy than the other state of the art techniques, and its training process takes less time than most of the traditional approaches presented in this paper.


Author(s):  
Rajat Khurana ◽  
Alok Kumar Singh Kushwaha

Background & Objective: Identification of human actions from video has gathered much attention in past few years. Most of the computer vision tasks such as Health Care Activity Detection, Suspicious Activity detection, Human Computer Interactions etc. are based on the principle of activity detection. Automatic labelling of activity from videos frames is known as activity detection. Motivation of this work is to use most out of the data generated from sensors and use them for recognition of classes. Recognition of actions from videos sequences is a growing field with the upcoming trends of deep neural networks. Automatic learning capability of Convolutional Neural Network (CNN) make them good choice as compared to traditional handcrafted based approaches. With the increasing demand of RGB-D sensors combination of RGB and depth data is in great demand. This work comprises of the use of dynamic images generated from RGB combined with depth map for action recognition purpose. We have experimented our approach on pre trained VGG-F model using MSR Daily activity dataset and UTD MHAD Dataset. We achieve state of the art results. To support our research, we have calculated different parameters apart from accuracy such as precision, F score, recall. Conclusion: Accordingly, the investigation confirms improvement in term of accuracy, precision, F-Score and Recall. The proposed model is 4 Stream model is prone to occlusion, used in real time and also the data from the RGB-D sensor is fully utilized.


Sign in / Sign up

Export Citation Format

Share Document