Author:
Moerland Thomas M.,Broekens Joost,Plaat Aske,Jonker Catholijn M.
Abstract
Sequential decision making, commonly formalized as optimization of a Markov Decision Process, is a key challenge in artificial intelligence. Two successful approaches to MDP optimization arereinforcement learningandplanning, which both largely have their own research communities. However, if both research fields solve the same problem, then we might be able to disentangle the common factors in their solution approaches. Therefore, this paper presents a unifying algorithmic framework for reinforcement learning and planning (FRAP), which identifies underlying dimensions on which MDP planning and learning algorithms have to decide. At the end of the paper, we compare a variety of well-known planning, model-free and model-based RL algorithms along these dimensions. Altogether, the framework may help provide deeper insight in the algorithmic design space of planning and reinforcement learning.
Reference166 articles.
1. Surprise-based intrinsic motivation for deep reinforcement learning;Achiam,2017
2. Solving the Rubik's cube with deep reinforcement learning and search;Agostinelli;Nat. Mach. Intell,2019
3. Binary decision diagrams;Akers;IEEE Trans. Comput,1978
4. “Revisiting regression in planning,”;Alcázar,2013
Cited by
7 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献