pose estimation
Recently Published Documents





2022 ◽  
Vol 134 ◽  
pp. 104089
Amin Assadzadeh ◽  
Mehrdad Arashpour ◽  
Ioannis Brilakis ◽  
Tuan Ngo ◽  
Erini Konstantinou

Sensors ◽  
2022 ◽  
Vol 22 (2) ◽  
pp. 632
Jie Li ◽  
Zhixing Wang ◽  
Bo Qi ◽  
Jianlin Zhang ◽  
Hu Yang

In this paper, a mutually enhanced modeling method (MEMe) is presented for human pose estimation, which focuses on enhancing lightweight model performance, but with low complexity. To obtain higher accuracy, a traditional model scale is largely expanded with heavy deployment difficulties. However, for a more lightweight model, there is a large performance gap compared to the former; thus, an urgent need for a way to fill it. Therefore, we propose a MEMe to reconstruct a lightweight baseline model, EffBase transferred intuitively from EfficientDet, into the efficient and effective pose (EEffPose) net, which contains three mutually enhanced modules: the Enhanced EffNet (EEffNet) backbone, the total fusion neck (TFNeck), and the final attention head (FAHead). Extensive experiments on COCO and MPII benchmarks show that our MEMe-based models reach state-of-the-art performances, with limited parameters. Specifically, in the same conditions, our EEffPose-P0 with 256 × 192 can use only 8.98 M parameters to achieve 75.4 AP on the COCO val set, which outperforms HRNet-W48, but with only 14% of its parameters.

2022 ◽  
pp. 9-18
Luca Lonini ◽  
Yaejin Moon ◽  
Kyle Embry ◽  
R. James Cotton ◽  
Kelly McKenzie ◽  

Recent advancements in deep learning have produced significant progress in markerless human pose estimation, making it possible to estimate human kinematics from single camera videos without the need for reflective markers and specialized labs equipped with motion capture systems. Such algorithms have the potential to enable the quantification of clinical metrics from videos recorded with a handheld camera. Here we used DeepLabCut, an open-source framework for markerless pose estimation, to fine-tune a deep network to track 5 body keypoints (hip, knee, ankle, heel, and toe) in 82 below-waist videos of 8 patients with stroke performing overground walking during clinical assessments. We trained the pose estimation model by labeling the keypoints in 2 frames per video and then trained a convolutional neural network to estimate 5 clinically relevant gait parameters (cadence, double support time, swing time, stance time, and walking speed) from the trajectory of these keypoints. These results were then compared to those obtained from a clinical system for gait analysis (GAITRite®, CIR Systems). Absolute accuracy (mean error) and precision (standard deviation of error) for swing, stance, and double support time were within 0.04 ± 0.11 s; Pearson’s correlation with the reference system was moderate for swing times (<i>r</i> = 0.4–0.66), but stronger for stance and double support time (<i>r</i> = 0.93–0.95). Cadence mean error was −0.25 steps/min ± 3.9 steps/min (<i>r</i> = 0.97), while walking speed mean error was −0.02 ± 0.11 m/s (<i>r</i> = 0.92). These preliminary results suggest that single camera videos and pose estimation models based on deep networks could be used to quantify clinically relevant gait metrics in individuals poststroke, even while using assistive devices in uncontrolled environments. Such development opens the door to applications for gait analysis both inside and outside of clinical settings, without the need of sophisticated equipment.

Micromachines ◽  
2022 ◽  
Vol 13 (1) ◽  
pp. 126
Lei Zhang ◽  
Huiliang Shang ◽  
Yandan Lin

The 6D Pose estimation is a crux in many applications, such as visual perception, autonomous navigation, and spacecraft motion. For robotic grasping, the cluttered and self-occlusion scenarios bring new challenges to the this field. Currently, society uses CNNs to solve this problem. The CNN models will suffer high uncertainty caused by the environmental factors and the object itself. These models usually maintain a Gaussian distribution, which is not suitable for the underlying manifold structure of the pose. Many works decouple rotation from the translation and quantify rotational uncertainty. Only a few works pay attention to the uncertainty of the 6D pose. This work proposes a distribution that can capture the uncertainty of the 6D pose parameterized by the dual quaternions, meanwhile, the proposed distribution takes the periodic nature of the underlying structure into account. The presented results include the normalization constant computation and parameter estimation techniques of the distribution. This work shows the benefits of the proposed distribution, which provides a more realistic explanation for the uncertainty in the 6D pose and eliminates the drawback inherited from the planar rigid motion.

Xue Wang ◽  
Runyang Feng ◽  
Haoming Chen ◽  
Roger Zimmermann ◽  
Zhenguang Liu ◽  

Jielu Yan ◽  
MingLiang Zhou ◽  
Jinli Pan ◽  
Meng Yin ◽  
Bin Fang

3D human pose estimation describes estimating 3D articulation structure of a person from an image or a video. The technology has massive potential because it can enable tracking people and analyzing motion in real time. Recently, much research has been conducted to optimize human pose estimation, but few works have focused on reviewing 3D human pose estimation. In this paper, we offer a comprehensive survey of the state-of-the-art methods for 3D human pose estimation, referred to as pose estimation solutions, implementations on images or videos that contain different numbers of people and advanced 3D human pose estimation techniques. Furthermore, different kinds of algorithms are further subdivided into sub-categories and compared in light of different methodologies. To the best of our knowledge, this is the first such comprehensive survey of the recent progress of 3D human pose estimation and will hopefully facilitate the completion, refinement and applications of 3D human pose estimation.

Chen Zhongshan ◽  
Feng Xinning ◽  
Oscar Sanjuán Martínez ◽  
Rubén González Crespo

In human-computer interaction and virtual truth, hand pose estimation is essential. Public dataset experimental analysis Different biometric shows that a particular system creates low manual estimation errors and has a more significant opportunity for new hand pose estimation activity. Due to the fluctuations, self-occlusion, and specific modulations, the structure of hand photographs is quite tricky. Hence, this paper proposes a Hybrid approach based on machine learning (HABoML) to enhance the current competitiveness, performance experience, experimental hand shape, and key point estimation analysis. In terms of strengthening the ability to make better self-occlusion adjustments and special handshake and poses estimations, the machine learning algorithm is combined with a hybrid approach. The experiment results helped define a set of follow-up experiments for the proposed systems in this field, which had a high efficiency and performance level. The HABoML strategy decreased analysis precision by 9.33% and is a better solution.

Sign in / Sign up

Export Citation Format

Share Document