Towards Multi-Objective Object Push-Grasp Policy Based on Maximum Entropy Deep Reinforcement Learning under Sparse Rewards
Author:
Zhang Tengteng1, Mo Hongwei1
Affiliation:
1. College of Intelligent Systems Science and Engineering, Harbin Engineering University, Harbin 150001, China
Abstract
In unstructured environments, robots need to deal with a wide variety of objects with diverse shapes, and often, the instances of these objects are unknown. Traditional methods rely on training with large-scale labeled data, but in environments with continuous and high-dimensional state spaces, the data become sparse, leading to weak generalization ability of the trained models when transferred to real-world applications. To address this challenge, we present an innovative maximum entropy Deep Q-Network (ME-DQN), which leverages an attention mechanism. The framework solves complex and sparse reward tasks through probabilistic reasoning while eliminating the trouble of adjusting hyper-parameters. This approach aims to merge the robust feature extraction capabilities of Fully Convolutional Networks (FCNs) with the efficient feature selection of the attention mechanism across diverse task scenarios. By integrating an advantage function with the reasoning and decision-making of deep reinforcement learning, ME-DQN propels the frontier of robotic grasping and expands the boundaries of intelligent perception and grasping decision-making in unstructured environments. Our simulations demonstrate a remarkable grasping success rate of 91.6%, while maintaining excellent generalization performance in the real world.
Reference34 articles.
1. Zhang, H., Lan, X., Bai, S., Zhou, X., Tian, Z., and Zheng, N. (2019, January 3–8). ROI-based Robotic Grasp Detection for Object Overlapping Scenes. Proceedings of the 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Macau, China. 2. Zhou, X., Lan, X., Zhang, H., Tian, Z., Zhang, Y., and Zheng, N. (2018, January 1–5). Fully Convolutional Grasp Detection Network with Oriented Anchor Box. Proceedings of the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain. 3. Chen, T., Shenoy, A., Kolinko, A., Shah, S., and Sun, Y. (October, January 7). Multi-Object Grasping—Estimating the Number of Objects in a Robotic Grasp. Proceedings of the 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Prague, Czech Republic. 4. Multimodal Data-Driven Robot Control for Human–Robot Collaborative Assembly;Liu;ASME. J. Manuf. Sci. Eng. May,2022 5. Valencia, D., Jia, J., Hayashi, A., Lecchi, M., Terezakis, R., Gee, T., Liarokapis, M., MacDonald, B.A., and Williams, H. (June, January 29). Comparison of Model-Based and Model-Free Reinforcement Learning for Real-World Dexterous Robotic Manipulation Tasks. Proceedings of the 2023 IEEE International Conference on Robotics and Automation (ICRA), London, UK.
|
|