action classification
Recently Published Documents


TOTAL DOCUMENTS

237
(FIVE YEARS 47)

H-INDEX

21
(FIVE YEARS 4)

Author(s):  
Mohammad Farhad Bulbul ◽  
Saiful Islam ◽  
Zannatul Azme ◽  
Preksha Pareek ◽  
Md. Humaun Kabir ◽  
...  

Sensors ◽  
2022 ◽  
Vol 22 (2) ◽  
pp. 599
Author(s):  
Yongsheng Li ◽  
Tengfei Tu ◽  
Hua Zhang ◽  
Jishuai Li ◽  
Zhengping Jin ◽  
...  

In the field of video action classification, existing network frameworks often only use video frames as input. When the object involved in the action does not appear in a prominent position in the video frame, the network cannot accurately classify it. We introduce a new neural network structure that uses sound to assist in processing such tasks. The original sound wave is converted into sound texture as the input of the network. Furthermore, in order to use the rich modal information (images and sound) in the video, we designed and used a two-stream frame. In this work, we assume that sound data can be used to solve motion recognition tasks. To demonstrate this, we designed a neural network based on sound texture to perform video action classification tasks. Then, we fuse this network with a deep neural network that uses continuous video frames to construct a two-stream network, which is called A-IN. Finally, in the kinetics dataset, we use our proposed A-IN to compare with the image-only network. The experimental results show that the recognition accuracy of the two-stream neural network model with uesed sound data features is increased by 7.6% compared with the network using video frames. This proves that the rational use of the rich information in the video can improve the classification effect.


Sensors ◽  
2021 ◽  
Vol 21 (24) ◽  
pp. 8409
Author(s):  
Rajesh Amerineni ◽  
Lalit Gupta ◽  
Nathan Steadman ◽  
Keshwyn Annauth ◽  
Charles Burr ◽  
...  

We introduce a set of input models for fusing information from ensembles of wearable sensors supporting human performance and telemedicine. Veracity is demonstrated in action classification related to sport, specifically strikes in boxing and taekwondo. Four input models, formulated to be compatible with a broad range of classifiers, are introduced and two diverse classifiers, dynamic time warping (DTW) and convolutional neural networks (CNNs) are implemented in conjunction with the input models. Seven classification models fusing information at the input-level, output-level, and a combination of both are formulated. Action classification for 18 boxing punches and 24 taekwondo kicks demonstrate our fusion classifiers outperform the best DTW and CNN uni-axial classifiers. Furthermore, although DTW is ostensibly an ideal choice for human movements experiencing non-linear variations, our results demonstrate deep learning fusion classifiers outperform DTW. This is a novel finding given that CNNs are normally designed for multi-dimensional data and do not specifically compensate for non-linear variations within signal classes. The generalized formulation enables subject-specific movement classification in a feature-blind fashion with trivial computational expense for trained CNNs. A commercial boxing system, ‘Corner’, has been produced for real-world mass-market use based on this investigation providing a basis for future telemedicine translation.


2021 ◽  
Author(s):  
Pierre-Etienne Martin ◽  
Jenny Benois-Pineau ◽  
Renaud Péteri ◽  
Julien Morlier

2021 ◽  
Author(s):  
Yi Tan ◽  
Yanbin Hao ◽  
Xiangnan He ◽  
Yinwei Wei ◽  
Xun Yang

2021 ◽  
Author(s):  
Jiawei Ma ◽  
Xiaoyu Tao ◽  
Jianxing Ma ◽  
Xiaopeng Hong ◽  
Yihong Gong

2021 ◽  
Vol 11 (18) ◽  
pp. 8633
Author(s):  
Katarzyna Gościewska ◽  
Dariusz Frejlichowski

This paper presents an action recognition approach based on shape and action descriptors that is aimed at the classification of physical exercises under partial occlusion. Regular physical activity in adults can be seen as a form of non-communicable diseases prevention, and may be aided by digital solutions that encourages individuals to increase their activity level. The application scenario includes workouts in front of the camera, where either the lower or upper part of the camera’s field of view is occluded. The proposed approach uses various features extracted from sequences of binary silhouettes, namely centroid trajectory, shape descriptors based on the Minimum Bounding Rectangle, action representation based on the Fourier transform and leave-one-out cross-validation for classification. Several experiments combining various parameters and shape features are performed. Despite the presence of occlusion, it was possible to obtain about 90% accuracy for several action classes, with the use of elongation values observed over time and centroid trajectory.


Sign in / Sign up

Export Citation Format

Share Document