scholarly journals Joint Dynamic Pose Image and Space Time Reversal for Human Action Recognition from Videos

Author(s):  
Mengyuan Liu ◽  
Fanyang Meng ◽  
Chen Chen ◽  
Songtao Wu

Human action recognition aims to classify a given video according to which type of action it contains. Disturbance brought by clutter background and unrelated motions makes the task challenging for video frame-based methods. To solve this problem, this paper takes advantage of pose estimation to enhance the performances of video frame features. First, we present a pose feature called dynamic pose image (DPI), which describes human action as the aggregation of a sequence of joint estimation maps. Different from traditional pose features using sole joints, DPI suffers less from disturbance and provides richer information about human body shape and movements. Second, we present attention-based dynamic texture images (att-DTIs) as pose-guided video frame feature. Specifically, a video is treated as a space-time volume, and DTIs are obtained by observing the volume from different views. To alleviate the effect of disturbance on DTIs, we accumulate joint estimation maps as attention map, and extend DTIs to attention-based DTIs (att-DTIs). Finally, we fuse DPI and att-DTIs with multi-stream deep neural networks and late fusion scheme for action recognition. Experiments on NTU RGB+D, UTD-MHAD, and Penn-Action datasets show the effectiveness of DPI and att-DTIs, as well as the complementary property between them.

2017 ◽  
Vol 11 (7) ◽  
pp. 530-540 ◽  
Author(s):  
Bassem Seddik ◽  
Sami Gazzah ◽  
Najoua Essoukri Ben Amara

2014 ◽  
Vol 36 ◽  
pp. 221-227 ◽  
Author(s):  
Antonio W. Vieira ◽  
Erickson R. Nascimento ◽  
Gabriel L. Oliveira ◽  
Zicheng Liu ◽  
Mario F.M. Campos

Author(s):  
Maxime Devanne ◽  
Hazem Wannous ◽  
Stefano Berretti ◽  
Pietro Pala ◽  
Mohamed Daoudi ◽  
...  

IEEE Access ◽  
2018 ◽  
Vol 6 ◽  
pp. 17913-17922 ◽  
Author(s):  
Lei Wang ◽  
Yangyang Xu ◽  
Jun Cheng ◽  
Haiying Xia ◽  
Jianqin Yin ◽  
...  

Author(s):  
Prof. Rajeshwari. J. Kodulkar

Abstract: In deep neural networks, human action detection is one of the most demanding and complex tasks. Human gesture recognition is the same as human action recognition. Gesture is defined as a series of bodily motions that communicate a message. Gestures are a more natural and preferable way for humans to engage with computers, thereby bridging the gap between humans and robots. The finest communication platform for the deaf and dumb is human action recognition. We propose in this work to create a system for hand gesture identification that recognizes hand movements, hand characteristics such as peak calculation and angle calculation, and then converts gesture photos into text. Index Terms: Human action recognition, Deaf and dumb, CNN.


Sign in / Sign up

Export Citation Format

Share Document