Modeling 3D Convolution Architecture for Actions Recognition

Author(s):  
Bogdan Alexandru Radulescu ◽  
Victorita Radulescu

Abstract Action recognition infrastructure can be applied anywhere behavior analysis is required and represents presently a domain of maximum actuality in security and surveillance. The model based on 3D Convolutions is a middle ground between simple key-frame approaches based on 2D convolutions, and other more complex approaches based on Recurrent Neural Networks. Behavior analysis represents a domain greatly improved by action recognition. By placing human actions in different categories it is possible to extract statistics regarding a person’s behavior, characteristics, abilities and preferences which can be processed later by specialized personnel, depending on the selected domain. The proposed model follows simple 3D convolution architecture. Hidden layers are composed of a convolution operation, an activation function and, sometimes, a pooling layer. Leaky ReLU was used as activation function to alleviate the problem of vanishing gradients. Batch Normalization is a technique used for scaling and adjusting the output of an activation layer, and it has been used to reduce over-fitting and decrease the training time. The 3D Convolution structure has the advantage of learning spatio-temporal features, because the convolution is applied over a sequence of frames. In the present paper is presented a proposed 3D convolution model that has average results, with an accuracy of approximately 55% on the NTU RGB+D dataset.

2020 ◽  
Vol 10 (15) ◽  
pp. 5326
Author(s):  
Xiaolei Diao ◽  
Xiaoqiang Li ◽  
Chen Huang

The same action takes different time in different cases. This difference will affect the accuracy of action recognition to a certain extent. We propose an end-to-end deep neural network called “Multi-Term Attention Networks” (MTANs), which solves the above problem by extracting temporal features with different time scales. The network consists of a Multi-Term Attention Recurrent Neural Network (MTA-RNN) and a Spatio-Temporal Convolutional Neural Network (ST-CNN). In MTA-RNN, a method for fusing multi-term temporal features are proposed to extract the temporal dependence of different time scales, and the weighted fusion temporal feature is recalibrated by the attention mechanism. Ablation research proves that this network has powerful spatio-temporal dynamic modeling capabilities for actions with different time scales. We perform extensive experiments on four challenging benchmark datasets, including the NTU RGB+D dataset, UT-Kinect dataset, Northwestern-UCLA dataset, and UWA3DII dataset. Our method achieves better results than the state-of-the-art benchmarks, which demonstrates the effectiveness of MTANs.


2018 ◽  
Vol 7 (4) ◽  
pp. 2153
Author(s):  
P A. Dhulekar ◽  
S T. Gandhe

In modern years large extent of the work has been carried out to recognize human actions perhaps because of its wide range of applications in the field of surveillance, human-machine interaction and video analysis. Several methods were proposed by researchers to resolve action recognition challenges such as variations in viewpoints, occlusion, cluttered backgrounds and camera motion. To address these challenges, we propose a novel method comprise of features extraction using histogram of oriented gradients (HOG), and their classification using k-nearest neighbor (k-NN) and support vector machine (SVM). Six different experimentations were carried out on the basis of hybrid combinations of feature extractors and classifiers. Two gold standard datasets; KTH and Weizmann were used for training and testing purpose. The quantitative parameters such as recognition accuracy, training time and prediction speed were used for evaluation. To validate the applicability of proposed algorithm, its performance has been compared with spatio-temporal interest points (STIP) technique which was proposed as state of art method in the domain. 


2014 ◽  
Vol 10 (1) ◽  
pp. 199-206 ◽  
Author(s):  
Lishen Pei ◽  
Mao Ye ◽  
Xuezhuan Zhao ◽  
Tao Xiang ◽  
Tao Li

Sign in / Sign up

Export Citation Format

Share Document