Your browser doesn't support javascript.
loading
DON6D: a decoupled one-stage network for 6D pose estimation.
Wang, Zheng; Tu, Hangyao; Qian, Yutong; Zhao, Yanwei.
Afiliação
  • Wang Z; School of Computer and Computational Sciences, Hangzhou City University, Hangzhou, 310015, China.
  • Tu H; School of Computer and Computational Sciences, Hangzhou City University, Hangzhou, 310015, China. lewieyao@yeah.net.
  • Qian Y; School of Computer Science and Technology, Zhejiang University of Technology, Hangzhou, 310023, China.
  • Zhao Y; School of Engineering, Hangzhou City University, Hangzhou, 310015, China.
Sci Rep ; 14(1): 8410, 2024 Apr 10.
Article em En | MEDLINE | ID: mdl-38600244
ABSTRACT
The six-dimensional (6D) pose object estimation is a key task in robotic manipulation and grasping scenes. Many existing two-stage solutions with a slow inference speed require extra refinement to handle the challenges of variations in lighting, sensor noise, object occlusion, and truncation. To address these challenges, this work proposes a decoupled one-stage network (DON6D) model for 6D pose estimation that improves inference speed on the premise of maintaining accuracy. Particularly, since the RGB images are aligned with the RGB-D images, the proposed DON6D first uses a two-dimensional detection network to locate the interested objects in RGB-D images. Then, a module of feature extraction and fusion is used to extract color and geometric features fully. Further, dual data augmentation is performed to enhance the generalization ability of the proposed model. Finally, the features are fused, and an attention residual encoder-decoder, which can improve the pose estimation performance to obtain an accurate 6D pose, is introduced. The proposed DON6D model is evaluated on the LINEMOD and YCB-Video datasets. The results demonstrate that the proposed DON6D is superior to several state-of-the-art methods regarding the ADD(-S) and ADD(-S) AUC metrics.
Palavras-chave

Texto completo: 1 Base de dados: MEDLINE Idioma: En Ano de publicação: 2024 Tipo de documento: Article

Texto completo: 1 Base de dados: MEDLINE Idioma: En Ano de publicação: 2024 Tipo de documento: Article