Affiliation:
1. College of Electrical Engineering , Northwest Minzu University, Lanzhou 730030, China
2. Gansu Engineering Research Center for Eco-Environmental Intelligent Networking, Lanzhou 730030, China
Abstract
Grasp detection serves as the fundamental element for achieving successful grasping in robotic systems. The encoder–decoder structure has become widely adopted as the foundational architecture for grasp detection networks due to its inherent advantages of speed and accuracy. However, traditional network structures fail to effectively extract the essential features required for accurate grasping poses and neglect to eliminate the checkerboard artifacts caused by inversion convolution during decoding. Aiming at overcoming these challenges, we propose a novel generative grasp detection network (LGAR-Net2). A transposed convolution layer is employed to replace the bilinear interpolation layer in the decoder to remove the issue of uneven overlapping and consequently eliminate checkerboard artifacts. In addition, a loss-guided collaborative attention block (LGCA), which combines attention blocks with spatial pyramid blocks to enhance the attention to important regions of the image, is constructed to enhance the accuracy of information extraction. Validated on the Cornell public dataset using RGB images as the input, LGAR-Net2 achieves an accuracy of 97.7%, an improvement of 1.1% over the baseline network, and processes a single RGB image in just 15 ms.
Funder
Fundamental Research Funds for the Central Universities of Northwest Minzu University
Gansu Province Demonstration major of Innovation and Entrepreneurship Education for Colleges and Universities
National Natural Science Foundation of China
Reference25 articles.
1. Ni, P., Zhang, W., Zhu, X., and Cao, Q. (August, January 31). PointNet++ grasping: Learning an end-to-end spatial grasp generation algorithm from sparse point clouds. Proceedings of the 2020 IEEE International Conference on Robotics and Automation (ICRA), Paris, France.
2. Jiang, Y., Moseson, S., and Saxena, A. (2011, January 9–13). Efficient grasping from RGBD images: Learning using a new rectangle representation. Proceedings of the 2011 IEEE International Conference on Robotics and Automation, Shanghai, China.
3. Deep learning for detecting robotic grasps;Lenz;Int. J. Robot. Res.,2015
4. Redmon, J., and Angelova, A. (2015, January 26–30). Real-time grasp detection using convolutional neural networks. Proceedings of the 2015 IEEE International Conference on Robotics and Automation (ICRA), Seattle, WA, USA.
5. Learning robust, real-time, reactive robotic grasping;Morrison;Int. J. Robot. Res.,2020