6D motion gesture recognition using spatio-temporal features

Author(s):  
Mingyu Chen ◽  
Ghassan AlRegib ◽  
Biing-Hwang Juang
2020 ◽  
Vol 10 (11) ◽  
pp. 3680 ◽  
Author(s):  
Chunyong Ma ◽  
Shengsheng Zhang ◽  
Anni Wang ◽  
Yongyang Qi ◽  
Ge Chen

Dynamic hand gesture recognition based on one-shot learning requires full assimilation of the motion features from a few annotated data. However, how to effectively extract the spatio-temporal features of the hand gestures remains a challenging issue. This paper proposes a skeleton-based dynamic hand gesture recognition using an enhanced network (GREN) based on one-shot learning by improving the memory-augmented neural network, which can rapidly assimilate the motion features of dynamic hand gestures. Besides, the network effectively combines and stores the shared features between dissimilar classes, which lowers the prediction error caused by the unnecessary hyper-parameters updating, and improves the recognition accuracy with the increase of categories. In this paper, the public dynamic hand gesture database (DHGD) is used for the experimental comparison of the state-of-the-art performance of the GREN network, and although only 30% of the dataset was used for training, the accuracy of skeleton-based dynamic hand gesture recognition reached 82.29% based on one-shot learning. Experiments with the Microsoft Research Asia (MSRA) hand gesture dataset verified the robustness of the GREN network. The experimental results demonstrate that the GREN network is feasible for skeleton-based dynamic hand gesture recognition based on one-shot learning.


2021 ◽  
Author(s):  
Jianxiao Xie ◽  
Wei Ye ◽  
Kai Xu

Abstract Internet of Things (IoT) expects to incorporate massive machine-type (MCT) devices, such as vehicles, sensors, and wearable devices, which brings a large number of application tasks that need to be processed. Additionally, data collected from various devices needs to be executed and processed in a timely, reliable, and efficient manner. Gesture recognition has enabled IoT applications such as human-computer interaction and virtual reality. In this work, we propose a cross-domain device-free gesture recognition (DFGR) model, that exploits 3D-CNN to obtain spatiotemporal features in Wi-Fi sensing. To adapt the sensing data to the 3D model, we carry out 3D data segment and supplement in addition to signal denoising and time-frequency transformation. We demonstrate that our proposed model outperforms the state-of-the-art method in the application of DFGR even cross 3 domain factors simultaneously, and is easy to converge and convenient for training with a less complicated hierarchical structure.


2021 ◽  
Author(s):  
Monir Torabian ◽  
Hossein Pourghassem ◽  
Homayoun Mahdavi-Nasab

2021 ◽  
pp. 115472
Author(s):  
Parameshwaran Ramalingam ◽  
Lakshminarayanan Gopalakrishnan ◽  
Manikandan Ramachandran ◽  
Rizwan Patan

2016 ◽  
Vol 12 ◽  
pp. P1115-P1115
Author(s):  
Vera Niederkofler ◽  
Christina Hoeller ◽  
Joerg Neddens ◽  
Ewald Auer ◽  
Heinrich Roemer ◽  
...  

2021 ◽  
Vol 12 (6) ◽  
pp. 1-23
Author(s):  
Shuo Tao ◽  
Jingang Jiang ◽  
Defu Lian ◽  
Kai Zheng ◽  
Enhong Chen

Mobility prediction plays an important role in a wide range of location-based applications and services. However, there are three problems in the existing literature: (1) explicit high-order interactions of spatio-temporal features are not systemically modeled; (2) most existing algorithms place attention mechanisms on top of recurrent network, so they can not allow for full parallelism and are inferior to self-attention for capturing long-range dependence; (3) most literature does not make good use of long-term historical information and do not effectively model the long-term periodicity of users. To this end, we propose MoveNet and RLMoveNet. MoveNet is a self-attention-based sequential model, predicting each user’s next destination based on her most recent visits and historical trajectory. MoveNet first introduces a cross-based learning framework for modeling feature interactions. With self-attention on both the most recent visits and historical trajectory, MoveNet can use an attention mechanism to capture the user’s long-term regularity in a more efficient way. Based on MoveNet, to model long-term periodicity more effectively, we add the reinforcement learning layer and named RLMoveNet. RLMoveNet regards the human mobility prediction as a reinforcement learning problem, using the reinforcement learning layer as the regularization part to drive the model to pay attention to the behavior with periodic actions, which can help us make the algorithm more effective. We evaluate both of them with three real-world mobility datasets. MoveNet outperforms the state-of-the-art mobility predictor by around 10% in terms of accuracy, and simultaneously achieves faster convergence and over 4x training speedup. Moreover, RLMoveNet achieves higher prediction accuracy than MoveNet, which proves that modeling periodicity explicitly from the perspective of reinforcement learning is more effective.


Sign in / Sign up

Export Citation Format

Share Document