Author:
Sasaki Hikaru,Horiuchi Tadashi,Kato Satoru, , ,
Abstract
Deep Q-network (DQN) is one of the most famous methods of deep reinforcement learning. DQN approximates the action-value function using Convolutional Neural Network (CNN) and updates it using Q-learning. In this study, we applied DQN to robot behavior learning in a simulation environment. We constructed the simulation environment for a two-wheeled mobile robot using the robot simulation software, Webots. The mobile robot acquired good behavior such as avoiding walls and moving along a center line by learning from high-dimensional visual information supplied as input data. We propose a method that reuses the best target network so far when the learning performance suddenly falls. Moreover, we incorporate Profit Sharing method into DQN in order to accelerate learning. Through the simulation experiment, we confirmed that our method is effective.
Publisher
Fuji Technology Press Ltd.
Subject
Artificial Intelligence,Computer Vision and Pattern Recognition,Human-Computer Interaction
Reference18 articles.
1. V. Mnih et al., “Playing Atari with Deep Reinforcement Learning,” Proc. of NIPS 2013 Deep Learning Workshop, 2013.
2. V. Mnih et al., “Human-level Control through Deep Reinforcement Learning,” Nature, Vol.518, pp. 529-533, 2015.
3. Y. LeCun et al., “Gradient-based Learning applied to Document Recognition,” Proc. of the IEEE, Vol.86, No.11, pp. 2278-2324, 1998.
4. Y. Matsuo, “Expectation of Robot Field from Artificial Intelligence Field,” J. of the Robotics Society of Japan, Vol.35, No.3, pp. 2-7, 2017 (in Japanese).
5. E. Yong, “Inside the Eye: Nature’s Most Exquisite Creation,” National Geographic Magazine, Vol.22, No.2, 2016.
Cited by
6 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献