Category-Level 6-D Object Pose Estimation With Learnable Prior Embeddings for Robotic Grasping

Sheng Yu, Di Hua Zhai*, Jian Yin, Yuanqing Xia

*此作品的通讯作者

科研成果: 期刊稿件文章同行评审

摘要

Category-level object pose estimation is crucial for predicting the poses of unknown objects within known categories. While methods relying on category-level object pose estimation with category priors necessitate prior training on datasets to acquire object priors, approaches for category-level object pose estimation without category priors lack relevant geometric information. To address these challenges, this article introduces a category-level object pose estimation method, PENet, based on learnable priors. The method utilizes a learnable category prior embedding to represent prior features and proposes a transformer-based prior embedding deformation module to initially deform the prior embedding from a global perspective to match the actual target object. Additionally, it introduces a transformer-based correspondence module to establish correspondence between instances and priors from a global perspective in order to further align the deformed feature embedding with the scene point cloud features. Experimental results demonstrate that the proposed method surpasses existing methods, achieving state-of-the-art performance on the dataset. Furthermore, the generalization ability of the proposed method is evaluated by applying PENet to object pose estimation on the Wild6D dataset, where it outperforms all related methods. Finally, the application of PENet to robotic grasping experiments on a real UR3 robot results in a higher success rate compared to previous methods.

源语言英语
期刊IEEE Transactions on Industrial Electronics
DOI
出版状态已接受/待刊 - 2025
已对外发布

指纹

探究 'Category-Level 6-D Object Pose Estimation With Learnable Prior Embeddings for Robotic Grasping' 的科研主题。它们共同构成独一无二的指纹。

引用此