FingerPoseNet: A finger-level multitask learning network with residual feature sharing for 3D hand pose estimation

Tekie Tsegay Tewolde, Ali Asghar Manjotho, Prodip Kumar Sarker, Zhendong Niu*

*此作品的通讯作者

科研成果: 期刊稿件文章同行评审

摘要

Hand pose estimation approaches commonly rely on shared hand feature maps to regress the 3D locations of all hand joints. Subsequently, they struggle to enhance finger-level features which are invaluable in capturing joint-to-finger associations and articulations. To address this limitation, we propose a finger-level multitask learning network with residual feature sharing, named FingerPoseNet, for accurate 3D hand pose estimation from a depth image. FingerPoseNet comprises three stages: (a) a shared base feature map extraction backbone based on pre-trained ResNet-50; (b) a finger-level multitask learning stage that extracts and enhances feature maps for each finger and the palm; and (c) a multitask fusion layer for consolidating the estimation results obtained by each subtask. We exploit multitask learning by decoupling the hand pose estimation task into six subtasks dedicated to each finger and palm. Each subtask is responsible for subtask-specific feature extraction, enhancement, and 3D keypoint regression. To enhance subtask-specific features, we propose a residual feature-sharing approach scaled up to mine supplementary information from all subtasks. Experiments performed on five challenging public hand pose datasets, including ICVL, NYU, MSRA, Hands-2019-Task1, and HO3D-v3 demonstrate significant improvements in accuracy compared with state-of-the-art approaches.

源语言英语
文章编号107315
期刊Neural Networks
187
DOI
出版状态已出版 - 7月 2025

指纹

探究 'FingerPoseNet: A finger-level multitask learning network with residual feature sharing for 3D hand pose estimation' 的科研主题。它们共同构成独一无二的指纹。

引用此