Safe Reinforcement Learning for Arm Manipulation with Constrained Markov Decision Process

Author:

Adjei Patrick1ORCID,Tasfi Norman1,Gomez-Rosero Santiago1ORCID,Capretz Miriam A. M.1ORCID

Affiliation:

1. Electrical and Computer Engineering, Western University, London, ON N6A 3K7, Canada

Abstract

In the world of human–robot coexistence, ensuring safe interactions is crucial. Traditional logic-based methods often lack the intuition required for robots, particularly in complex environments where these methods fail to account for all possible scenarios. Reinforcement learning has shown promise in robotics due to its superior adaptability over traditional logic. However, the exploratory nature of reinforcement learning can jeopardize safety. This paper addresses the challenges in planning trajectories for robotic arm manipulators in dynamic environments. In addition, this paper highlights the pitfalls of multiple reward compositions that are susceptible to reward hacking. A novel method with a simplified reward and constraint formulation is proposed. This enables the robot arm to avoid a nonstationary obstacle that never resets, enhancing operational safety. The proposed approach combines scalarized expected returns with a constrained Markov decision process through a Lagrange multiplier, resulting in better performance. The scalarization component uses the indicator cost function value, directly sampled from the replay buffer, as an additional scaling factor. This method is particularly effective in dynamic environments where conditions change continually, as opposed to approaches relying solely on the expected cost scaled by a Lagrange multiplier.

Funder

Natural Sciences and Engineering Research Council

Publisher

MDPI AG

Reference43 articles.

1. Colgate, E., Bicchi, A., Peshkin, M.A., and Colgate, J.E. (2008). Springer Handbook of Robotics, Springer.

2. Beetz, M., Chatila, R., Hertzberg, J., and Pecora, F. (2016). AI Reasoning Methods for Robotics, Springer.

3. Deliberation for autonomous robots: A survey;Ingrand;Artif. Intell.,2017

4. Human-level control through deep reinforcement learning;Mnih;Nature,2015

5. Hasselt, H.V., Guez, A., and Silver, D. (2016, January 12–17). Deep Reinforcement Learning with Double Q-Learning. Proceedings of the Thirtieth AAAI Conference on Artificial Intelligence, AAAI’16, Phoenix, AZ, USA.

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3