Affiliation:
1. Department of Industrial Engineering, University of Florence, 50139 Florence, Italy
Abstract
Vision-based pose detection and grasping complex-shaped and thin objects are challenging tasks. We propose an architecture that integrates the Generative Grasping Convolutional Neural Network (GG-CNN) with depth recognition to identify a suitable grasp pose. First, we construct a training dataset with data augmentation to train a GG-CNN with only RGB images. Then, we extract a segment of the tool using a color segmentation method and use it to calculate an average depth. Additionally, we apply and evaluate different encoder–decoder models with a GG-CNN structure using the Intersection Over Union (IOU). Finally, we validate the proposed architecture by performing real-world grasping and pick-and-place experiments. Our framework achieves a success rate of over 85.6% for picking and placing seen surgical tools and 90% for unseen surgical tools. We collected a dataset of surgical tools and validated their pick and place with different GG-CNN architectures. In the future, we aim to expand the dataset of surgical tools and improve the accuracy of the GG-CNN.
Reference58 articles.
1. Morrison, D., Corke, P., and Leitner, J. (2018, January 26–30). Closing the Loop for Robotic Grasping: A Real-time, Generative Grasp Synthesis Approach. Proceedings of the Robotics: Science and Systems (RSS), Pittsburgh, PA, USA.
2. Learning robust, real-time, reactive robotic grasping;Morrison;Int. J. Robot. Res.,2020
3. Data-driven grasp synthesis—A survey;Bohg;IEEE Trans. Robot.,2013
4. An overview of 3D object grasp synthesis algorithms;Sahbani;Robot. Auton. Syst.,2012
5. Mousavian, A., Eppner, C., and Fox, D. (November, January 27). 6-Dof graspnet: Variational grasp generation for object manipulation. Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea.
Cited by
2 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献