pose recognition
Recently Published Documents


TOTAL DOCUMENTS

210
(FIVE YEARS 44)

H-INDEX

15
(FIVE YEARS 2)

Computers ◽  
2021 ◽  
Vol 11 (1) ◽  
pp. 2
Author(s):  
Srinivasan Raman ◽  
Rytis Maskeliūnas ◽  
Robertas Damaševičius

The analysis and perception of behavior has usually been a crucial task for researchers. The goal of this paper is to address the problem of recognition of animal poses, which has numerous applications in zoology, ecology, biology, and entertainment. We propose a methodology to recognize dog poses. The methodology includes the extraction of frames for labeling from videos and deep convolutional neural network (CNN) training for pose recognition. We employ a semi-supervised deep learning model of reinforcement. During training, we used a combination of restricted labeled data and a large amount of unlabeled data. Sequential CNN is also used for feature localization and to find the canine’s motions and posture for spatio-temporal analysis. To detect the canine’s features, we employ image frames to locate the annotations and estimate the dog posture. As a result of this process, we avoid starting from scratch with the feature model and reduce the need for a large dataset. We present the results of experiments on a dataset of more than 5000 images of dogs in different poses. We demonstrated the effectiveness of the proposed methodology for images of canine animals in various poses and behavior. The methodology implemented as a mobile app that can be used for animal tracking.


2021 ◽  
Vol 12 ◽  
Author(s):  
Chengming Ma ◽  
Qian Liu ◽  
Yaqi Dang

This paper provides an in-depth study and analysis of human artistic poses through intelligently enhanced multimodal artistic pose recognition. A complementary network model architecture of multimodal information based on motion energy proposed. The network exploits both the rich information of appearance features provided by RGB data and the depth information provided by depth data as well as the characteristics of robustness to luminance and observation angle. The multimodal fusion is accomplished by the complementary information characteristics of the two modalities. Moreover, to better model the long-range temporal structure while considering action classes with sub-action sharing phenomena, an energy-guided video segmentation method is employed. And in the feature fusion stage, a cross-modal cross-fusion approach is proposed, which enables the convolutional network to share local features of two modalities not only in the shallow layer but also to obtain the fusion of global features in the deep convolutional layer by connecting the feature maps of multiple convolutional layers. Firstly, the Kinect camera is used to acquire the color image data of the human body, the depth image data, and the 3D coordinate data of the skeletal points using the Open pose open-source framework. Then, the action automatically extracted from keyframes based on the distance between the hand and the head, and the relative distance features are extracted from the keyframes to describe the action, the local occupancy pattern features and HSV color space features are extracted to describe the object, and finally, the feature fusion is performed and the complex action recognition task is completed. To solve the consistency problem of virtual-reality fusion, the mapping relationship between hand joint point coordinates and the virtual scene is determined in the augmented reality scene, and the coordinate consistency model of natural hand and virtual model is established; finally, the real-time interaction between hand gesture and virtual model is realized, and the average correct rate of its hand gesture reaches 99.04%, which improves the robustness and real-time interaction of hand gesture recognition.


Sensors ◽  
2021 ◽  
Vol 21 (21) ◽  
pp. 7130
Author(s):  
Difei Xu ◽  
Xuelei Qi ◽  
Chen Li ◽  
Ziheng Sheng ◽  
Hailong Huang

The growing problem of aging has led to a social concern on how to take care of the elderly living alone. Many traditional methods based on visual cameras have been used in elder monitoring. However, these methods are difficult to be applied in daily life, limited by high storage space with the camera, low-speed information processing, sensitivity to lighting, the blind area in vision, and the possibility of revealing privacy. Therefore, wise information technology of the Med System based on the micro-Doppler effect and Ultra Wide Band (UWB) radar for human pose recognition in the elderly living alone is proposed to effectively identify and classify the human poses in static and moving conditions. In recognition processing, an improved PCA-LSTM approach is proposed by combing with the Principal Component Analysis (PCA) and Long Short Term Memory (LSTM) to integrate the micro-Doppler features and time sequence of the human body to classify and recognize the human postures. Moreover, the classification accuracy with different kernel functions in the Support Vector Machine (SVM) is also studied. In the real experiment, there are two healthy men and one woman (22–26 years old) selected to imitate the movements of the elderly and slowly perform five postures (from sitting to standing, from standing to sitting, walking in place, falling and boxing). The experimental results show that the resolution of the entire system for the five actions reaches 99.1% in the case of using Gaussian kernel function, so the proposed method is effective and the Gaussian kernel function is suitable for human pose recognition.


2021 ◽  
Author(s):  
Junyi Ma ◽  
Guangming Xiong ◽  
Jingyi Xu ◽  
Jiarui Song ◽  
Dong Sun

Author(s):  
Kuo-Chin Lin ◽  
Cheng-Wen Ko ◽  
Hui-Chun Hung ◽  
Nian-Shing Chen

Sign in / Sign up

Export Citation Format

Share Document