Constraint-Aware Policy for Compliant Manipulation
Author:
Saito Daichi1ORCID, Sasabuchi Kazuhiro2, Wake Naoki2ORCID, Kanehira Atsushi2, Takamatsu Jun2, Koike Hideki1, Ikeuchi Katsushi2
Affiliation:
1. School of Computing, Tokyo Institute of Technology, Tokyo 152-8550, Japan 2. Applied Robotics Research, Microsoft, Redmond, WA 98052, USA
Abstract
Robot manipulation in a physically constrained environment requires compliant manipulation. Compliant manipulation is a manipulation skill to adjust hand motion based on the force imposed by the environment. Recently, reinforcement learning (RL) has been applied to solve household operations involving compliant manipulation. However, previous RL methods have primarily focused on designing a policy for a specific operation that limits their applicability and requires separate training for every new operation. We propose a constraint-aware policy that is applicable to various unseen manipulations by grouping several manipulations together based on the type of physical constraint involved. The type of physical constraint determines the characteristic of the imposed force direction; thus, a generalized policy is trained in the environment and reward designed on the basis of this characteristic. This paper focuses on two types of physical constraints: prismatic and revolute joints. Experiments demonstrated that the same policy could successfully execute various compliant manipulation operations, both in the simulation and reality. We believe this study is the first step toward realizing a generalized household robot.
Subject
Artificial Intelligence,Control and Optimization,Mechanical Engineering
Reference40 articles.
1. Compliance and force control for computer controlled manipulators;Mason;IEEE Trans. Syst. Man Cybern.,1981 2. Yahya, A., Li, A., Kalakrishnan, M., Chebotar, Y., and Levine, S. (2017, January 24–28). Collective robot reinforcement learning with distributed asynchronous guided policy search. Proceedings of the 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Vancouver, BC, Canada. 3. Gu, S., Holly, E., Lillicrap, T., and Levine, S. (June, January 29). Deep reinforcement learning for robotic manipulation with asynchronous off-policy updates. Proceedings of the 2017 IEEE International Conference on Robotics and Automation (ICRA), Singapore. 4. Rajeswaran, A., Kumar, V., Gupta, A., Vezzani, G., Schulman, J., Todorov, E., and Levine, S. (2018, January 26–30). Learning Complex Dexterous Manipulation with Deep Reinforcement Learning and Demonstrations. Proceedings of the Robotics: Science and Systems (RSS), Pittsburgh, PA, USA. 5. Urakami, Y., Hodgkinson, A., Carlin, C., Leu, R., Rigazio, L., and Abbeel, P. (2019). Doorgym: A scalable door opening environment and baseline agent. arXiv.
|
|