Author:
Hessel Matteo,Modayil Joseph,Van Hasselt Hado,Schaul Tom,Ostrovski Georg,Dabney Will,Horgan Dan,Piot Bilal,Azar Mohammad,Silver David
Abstract
The deep reinforcement learning community has made several independent improvements to the DQN algorithm. However, it is unclear which of these extensions are complementary and can be fruitfully combined. This paper examines six extensions to the DQN algorithm and empirically studies their combination. Our experiments show that the combination provides state-of-the-art performance on the Atari 2600 benchmark, both in terms of data efficiency and final performance. We also provide results from a detailed ablation study that shows the contribution of each component to overall performance.
Publisher
Association for the Advancement of Artificial Intelligence (AAAI)
Cited by
402 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献