Author:
Graña Manuel,Fernandez-Gauna Borja,Lopez-Guede Jose Manuel
Abstract
AbstractReinforcement Learning (RL) as a paradigm aims to develop algorithms that allow to train an agent to optimally achieve a goal with minimal feedback information about the desired behavior, which is not precisely specified. Scalar rewards are returned to the agent as response to its actions endorsing or opposing them. RL algorithms have been successfully applied to robot control design. The extension of the RL paradigm to cope with the design of control systems for Multi-Component Robotic Systems (MCRS) poses new challenges, mainly related to coping with scaling up of complexity due to the exponential state space growth, coordination issues, and the propagation of rewards among agents. In this paper, we identify the main issues which offer opportunities to develop innovative solutions towards fully-scalable cooperative multi-agent systems.
Subject
Behavioral Neuroscience,Artificial Intelligence,Cognitive Neuroscience,Developmental Neuroscience,Human-Computer Interaction
Reference77 articles.
1. David W. Aha, Dennis Kibler, and Marc K. Albert. Instance-based learning algorithms. In Machine Learning, pages 37–66, 1991.
2. R. Aragues, J. Cortes, and C. Sagues. Distributed consensus algorithms for merging feature-based maps with limited communication. Robotics and Autonomous Systems, 59(3–4):163–180, 2011.
3. Andrew G. Barto. Using relative novelty to identify useful temporal abstractions in reinforcement learning. In In Procedings of the Twenty-First International Conference on Machine Learning, pages 751–758. ACM Press, 2004.
4. Hamid Berenji. Fuzzy reinforcement learning and dynamic programming. In Anca Ralescu, editor, Fuzy Logic in Artificial Inteligence, volume 847 of Lecture Notes in Computer Science, pages 1–9. Springer Berlin / Heidelberg, 1994.
5. H.R. Berenji. Fuzzy Q-learning for generalization of reinforcement learning. In IEEE Press, editor, Proc. of the Fifth IEEE International Conference on Fuzy Systems, volume 3, pages 2208–2214, 1996.
Cited by
1 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献