Refined Prior Guided Category-Level 6D Pose Estimation and Its Application on Robotic Grasping
-
Published:2024-09-07
Issue:17
Volume:14
Page:8009
-
ISSN:2076-3417
-
Container-title:Applied Sciences
-
language:en
-
Short-container-title:Applied Sciences
Author:
Sun Huimin1ORCID, Zhang Yilin1ORCID, Sun Honglin1ORCID, Hashimoto Kenji1ORCID
Affiliation:
1. Graduate School of Information, Production and Systems, Waseda University, Kitakyushu 808-0135, Japan
Abstract
Estimating the 6D pose and size of objects is crucial in the task of visual grasping for robotic arms. Most current algorithms still require the 3D CAD model of the target object to match with the detected points, and they are unable to predict the object’s size, which significantly limits the generalizability of these methods. In this paper, we introduce category priors and extract high-dimensional abstract features from both the observed point cloud and the prior to predict the deformation matrix of the reconstructed point cloud and the dense correspondence between the reconstructed and observed point clouds. Furthermore, we propose a staged geometric correction and dense correspondence refinement mechanism to enhance the accuracy of regression. In addition, a novel lightweight attention module is introduced to further integrate the extracted features and identify potential correlations between the observed point cloud and the category prior. Ultimately, the object’s translation, rotation, and size are obtained by mapping the reconstructed point cloud to a normalized canonical coordinate system. Through extensive experiments, we demonstrate that our algorithm outperforms existing methods in terms of performance and accuracy on commonly used benchmarks for this type of problem. Additionally, we implement the algorithm in robotic arm-grasping simulations, further validating its effectiveness.
Funder
Future Robotics Organization, Waseda University Humanoid Robotics Institute, Waseda University JSPS KAKENHI Waseda University Grant for Special Research Projects JST SPRING
Reference35 articles.
1. Kumra, S., Joshi, S., and Sahin, F. (2020–24, January 24). Antipodal robotic grasping using generative residual convolutional neural network. Proceedings of the 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Las Vegas, NV, USA. 2. Learning robust, real-time, reactive robotic grasping;Morrison;Int. J. Robot. Res.,2020 3. Sahin, C., Garcia-Hernando, G., Sock, J., and Kim, T.K. (2019). Instance-and category-level 6D object pose estimation. RGB-D Image Analysis and Processing, Springer. 4. Wang, C., Xu, D., Zhu, Y., Martín-Martín, R., Lu, C., Fei-Fei, L., and Savarese, S. (2019, January 15–20). Densefusion: 6D object pose estimation by iterative dense fusion. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA. 5. Tremblay, J., To, T., Sundaralingam, B., Xiang, Y., Fox, D., and Birchfield, S. (2018). Deep object pose estimation for semantic robotic grasping of household objects. arXiv.
|
|