Improving Object Grasp Performance via Transformer-Based Sparse Shape Completion
-
Published:2022-02-26
Issue:3
Volume:104
Page:
-
ISSN:0921-0296
-
Container-title:Journal of Intelligent & Robotic Systems
-
language:en
-
Short-container-title:J Intell Robot Syst
Author:
Chen WenkaiORCID, Liang HongzhuoORCID, Chen Zhaopeng, Sun Fuchun, Zhang Jianwei
Abstract
AbstractCurrently, robotic grasping methods based on sparse partial point clouds have attained excellent grasping performance on various objects. However, they often generate wrong grasping candidates due to the lack of geometric information on the object. In this work, we propose a novel and robust sparse shape completion model (TransSC). This model has a transformer-based encoder to explore more point-wise features and a manifold-based decoder to exploit more object details using a segmented partial point cloud as input. Quantitative experiments verify the effectiveness of the proposed shape completion network and demonstrate that our network outperforms existing methods. Besides, TransSC is integrated into a grasp evaluation network to generate a set of grasp candidates. The simulation experiment shows that TransSC improves the grasping generation result compared to the existing shape completion baselines. Furthermore, our robotic experiment shows that with TransSC, the robot is more successful in grasping objects of unknown numbers randomly placed on a support surface.
Funder
deutsches krebsforschungszentrum national natural science foundation of china h2020 european research council Universität Hamburg
Publisher
Springer Science and Business Media LLC
Subject
Electrical and Electronic Engineering,Artificial Intelligence,Industrial and Manufacturing Engineering,Mechanical Engineering,Control and Systems Engineering,Software
Reference36 articles.
1. Achlioptas, P., Diamanti, O., Mitliagkas, I., Guibas, L.: Learning representations and generative models for 3D point clouds. In: International Conference on Machine Learning, pp. 40–49. PMLR (2018) 2. Badrinarayanan, V., Kendall, A., Cipolla, R.: Segnet: a deep convolutional encoder-decoder architecture for image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 39(12), 2481–2495 (2017) 3. Bohg, J., Johnson-Roberson, M., León, B., Felip, J., Gratal, X., Bergström, N., Kragic, D., Morales, A.: Mind the gap-robotic grasping under incomplete observation. In: 2011 IEEE International Conference on Robotics and Automation, pp. 686–693. IEEE (2011) 4. Breyer, M., Chung, J.J., Ott, L., Siegwart, R., Nieto, J.: Volumetric grasping network: real-time 6 dof grasp detection in clutter. arXiv:2101.01132 (2021) 5. Chu, F.J., Xu, R., Vela, P.A.: Real-world multiobject, multigrasp detection. IEEE Robot. Autom. Lett. 3(4), 3355–3362 (2018)
Cited by
9 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
|
|