FAGD-Net: Feature-Augmented Grasp Detection Network Based on Efficient Multi-Scale Attention and Fusion Mechanisms
-
Published:2024-06-12
Issue:12
Volume:14
Page:5097
-
ISSN:2076-3417
-
Container-title:Applied Sciences
-
language:en
-
Short-container-title:Applied Sciences
Author:
Zhong Xungao12, Liu Xianghui1, Gong Tao1, Sun Yuan12, Hu Huosheng3ORCID, Liu Qiang4
Affiliation:
1. School of Electrical Engineering and Automation, Xiamen University of Technology, Xiamen 361024, China 2. Xiamen Key Laboratory of Frontier Electric Power Equipment and Intelligent Control, Xiamen 361024, China 3. School of Computer Science and Electronic Engineering, University of Essex, Colchester CO4 3SQ, UK 4. School of Engineering Mathematics and Technology, Faculty of Engineering, University of Bristol, Beacon House, Queens Rd, Bristol BS8 1QU, UK
Abstract
Grasping robots always confront challenges such as uncertainties in object size, orientation, and type, necessitating effective feature augmentation to improve grasping detection performance. However, many prior studies inadequately emphasize grasp-related features, resulting in suboptimal grasping performance. To address this limitation, this paper proposes a new grasping approach termed the Feature-Augmented Grasp Detection Network (FAGD-Net). The proposed network incorporates two modules designed to enhance spatial information features and multi-scale features. Firstly, we introduce the Residual Efficient Multi-Scale Attention (Res-EMA) module, which effectively adjusts the importance of feature channels while preserving precise spatial information within those channels. Additionally, we present a Feature Fusion Pyramidal Module (FFPM) that serves as an intermediary between the encoder and decoder, effectively addressing potential oversights or losses of grasp-related features as the encoder network deepens. As a result, FAGD-Net achieved advanced levels of grasping accuracy, with 98.9% and 96.5% on the Cornell and Jacquard datasets, respectively. The grasp detection model was deployed on a physical robot for real-world grasping experiments, where we conducted a series of trials in diverse scenarios. In these experiments, we randomly selected various unknown household items and adversarial objects. Remarkably, we achieved high success rates, with a 95.0% success rate for single-object household items, 93.3% for multi-object scenarios, and 91.0% for cluttered scenes.
Funder
National Natural Science Foundation of China Natural Science Foundation of Fujian Province Xiamen Natural Science Foundation
Reference42 articles.
1. Evolution strategies learning with variable impedance control for grasping under uncertainty;Hu;IEEE Trans. Ind. Electron.,2019 2. Adaptive Graph Convolutional Network with Adversarial Learning for Skeleton-Based Action Prediction;Li;IEEE Trans. Cogn. Dev. Syst.,2022 3. Solowjow, E., Ugalde, I., Shahapurkar, Y., Aparicio, J., Mahler, J., Satish, V., Goldberg, K., and Claussen, H. (2020, January 20–21). Industrial Robot Grasping with Deep Learning using a Programmable Logic Controller (PLC). Proceedings of the 2020 IEEE 16th International Conference on Automation Science and Engineering (CASE), Hong Kong, China. 4. Learning robust, real-time, reactive robotic grasping;Morrison;Int. J. Robot. Res.,2020 5. Mahler, J., Liang, J., Niyaz, S., Laskey, M., Doan, R., Liu, X., Ojea, J.A., and Goldberg, K. (2017). Dex-net 2.0: Deep learning to plan robust grasps with synthetic point clouds and analytic grasp metrics. arXiv.
|
|