HandFormer: Hand pose reconstructing from a single RGB image

Zixun Jiao, Xihan Wang, Jingcao Li, Rongxin Gao, Miao He, Jiao Liang, Zhaoqiang Xia, Quanli Gao

科研成果: 期刊稿件文章同行评审

1 引用 (Scopus)

摘要

We propose a multi-task progressive Transformer framework to reconstruct hand poses from a single RGB image to address challenges such as hand occlusion hand distraction, and hand shape bias. Our proposed framework comprises three key components: the feature extraction branch, palm segmentation branch, and parameter prediction branch. The feature extraction branch initially employs the progressive Transformer to extract multi-scale features from the input image. Subsequently, these multi-scale features are fed into a multi-layer perceptron layer (MLP) for acquiring palm alignment features. We employ an efficient fusion module to enhance the parameter prediction further features to integrate the palm alignment features with the backbone features. A dense hand model is generated using a pre-computed articulated mesh deformed hand model. We evaluate the performance of our proposed method on STEREO, FreiHAND, and HO3D datasets separately. The experimental results demonstrate that our approach achieves 3D mean error metrics of 10.92 mm, 12.33 mm and 9.6 mm for the respective datasets.

源语言英语
页(从-至)155-164
页数10
期刊Pattern Recognition Letters
183
DOI
出版状态已出版 - 7月 2024

指纹

探究 'HandFormer: Hand pose reconstructing from a single RGB image' 的科研主题。它们共同构成独一无二的指纹。

引用此