Affiliation:
1. Washington State University
Abstract
Reinforcement learning has had many successes, but in practice it often requires significant amounts of data to learn high-performing policies. One common way to improve learning is to allow a trained (source) agent to assist a new (target) agent. The goals in this setting are to 1) improve the target agent's performance, relative to learning unaided, and 2) allow the target agent to outperform the source agent. Our approach leverages source agent demonstrations, removing any requirements on the source agent's learning algorithm or representation. The target agent then estimates the source agent's policy and improves upon it. The key contribution of this work is to show that leveraging the target agent's uncertainty in the source agent's policy can significantly improve learning in two complex simulated domains, Keepaway and Mario.
Publisher
International Joint Conferences on Artificial Intelligence Organization
Cited by
14 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
1. Safe Exploration in Reinforcement Learning for Learning from Human Experts;2023 IEEE International Conference on Artificial Intelligence, Blockchain, and Internet of Things (AIBThings);2023-09-16
2. Reinforcement learning from expert demonstrations with application to redundant robot control;Engineering Applications of Artificial Intelligence;2023-03
3. Learning by reusing previous advice: a memory-based teacher–student framework;Autonomous Agents and Multi-Agent Systems;2022-12-29
4. Redundant robot control with learning from expert demonstrations;2022 IEEE Symposium Series on Computational Intelligence (SSCI);2022-12-04
5. Learning from Unreliable Human Action Advice in Interactive Reinforcement Learning;2022 IEEE-RAS 21st International Conference on Humanoid Robots (Humanoids);2022-11-28