Quadrotor Path Following and Reactive Obstacle Avoidance with Deep Reinforcement Learning
-
Published:2021-11-10
Issue:4
Volume:103
Page:
-
ISSN:0921-0296
-
Container-title:Journal of Intelligent & Robotic Systems
-
language:en
-
Short-container-title:J Intell Robot Syst
Author:
Rubí BartomeuORCID, Morcego Bernardo, Pérez Ramon
Abstract
AbstractA deep reinforcement learning approach for solving the quadrotor path following and obstacle avoidance problem is proposed in this paper. The problem is solved with two agents: one for the path following task and another one for the obstacle avoidance task. A novel structure is proposed, where the action computed by the obstacle avoidance agent becomes the state of the path following agent. Compared to traditional deep reinforcement learning approaches, the proposed method allows to interpret the training process outcomes, is faster and can be safely trained on the real quadrotor. Both agents implement the Deep Deterministic Policy Gradient algorithm. The path following agent was developed in a previous work. The obstacle avoidance agent uses the information provided by a low-cost LIDAR to detect obstacles around the vehicle. Since LIDAR has a narrow field-of-view, an approach for providing the agent with a memory of the previously seen obstacles is developed. A detailed description of the process of defining the state vector, the reward function and the action of this agent is given. The agents are programmed in python/tensorflow and are trained and tested in the RotorS/gazebo platform. Simulations results prove the validity of the proposed approach.
Funder
Universitat Politècnica de Catalunya
Publisher
Springer Science and Business Media LLC
Subject
Electrical and Electronic Engineering,Artificial Intelligence,Industrial and Manufacturing Engineering,Mechanical Engineering,Control and Systems Engineering,Software
Reference35 articles.
1. Caicedo, J.C., Lazebnik, S.: Active object localization with deep reinforcement learning, in. In: 2015 IEEE International Conference on Computer Vision (ICCV), pp. 2488–2496 (2015) 2. Silver, D., Huang, A., Maddison, C.J., Guez, A., Sifre, L., van den Driessche, G., Schrittwieser, J., Antonoglou, I., Panneershelvam, V., Lanctot, M., Dieleman, S., Grewe, D., Nham, J., Kalchbrenner, N., Sutskever, I., Lillicrap, T.P., Leach, M., Kavukcuoglu, K., Graepel, T., Hassabis, D.: Mastering the game of go with deep neural networks and tree search. Nature 529, 484–489 (2016) 3. Yu, R., Shi, Z., Huang, C., Li, T., Ma, Q.: Deep reinforcement learning based optimal trajectory tracking control of autonomous underwater vehicle. In: 2017 36th Chinese Control Conference (CCC), pp. 4958–4965 (2017) 4. Tuyen, L.P., Chung, T.: Controlling bicycle using deep deterministic policy gradient algorithm. In: 2017 14TH International Conference on Ubiquitous Robots and Ambien Intelligence (Urai), pp. 413–417 (2017) 5. Li, L., Lv, Y., Wang, F.: Traffic signal timing via deep reinforcement learning. IEEE/CAA J. Automatica Sinica 3(3), 247–254 (2016)
Cited by
17 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
|
|