Reconstructing Humans with a Biomechanically Accurate Skeleton
Abstract
In this paper, we introduce a method for reconstructing 3D humans from a single image using a biomechanically accurate skeleton model. To achieve this, we train a transformer that takes an image as input and estimates the parameters of the model. Due to the lack of training data for this task, we build a pipeline to produce pseudo ground truth model parameters for single images and implement a training procedure that iteratively refines these pseudo labels. Compared to state-of-the-art methods for 3D human mesh recovery, our model achieves competitive performance on standard benchmarks, while it significantly outperforms them in settings with extreme 3D poses and viewpoints. Additionally, we show that previous reconstruction methods frequently violate joint angle limits, leading to unnatural rotations. In contrast, our approach leverages the biomechanically plausible degrees of freedom making more realistic joint rotation estimates. We validate our approach across multiple human pose estimation benchmarks. We make the code, models and data available at: https://isshikihugh.github.io/HSMR/
Community
- We present HSMR (Human Skeleton and Mesh Recovery), the first end-to-end approach to recover SKEL parameters from single image.
- We show how to create a dataset with pseudo ground truth to train a model for other human body models.
- We demonstrate that HSMR shows robustness in extreme poses and viewpoints, providing biomechanically accurate human pose estimation, while still matches the performance of the most closely related state-of-the-art method that regresses SMPL parameters.
- We reveal the limitations of previous methods regressing SMPL parameters, and show how they tend to predict unnatural rotations for the body joints, leading to biomechanically inaccurate results.
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- LHM: Large Animatable Human Reconstruction Model from a Single Image in Seconds (2025)
- Realistic Clothed Human and Object Joint Reconstruction from a Single Image (2025)
- Fish2Mesh Transformer: 3D Human Mesh Recovery from Egocentric Vision (2025)
- FRESA:Feedforward Reconstruction of Personalized Skinned Avatars from Few Images (2025)
- AniGaussian: Animatable Gaussian Avatar with Pose-guided Deformation (2025)
- ReJSHand: Efficient Real-Time Hand Pose Estimation and Mesh Reconstruction Using Refined Joint and Skeleton Features (2025)
- CHROME: Clothed Human Reconstruction with Occlusion-Resilience and Multiview-Consistency from a Single Image (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 1
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 1
Collections including this paper 0
No Collection including this paper