Abstract
AbstractReinforcement learning has previously been applied to the problem of controlling a perched landing manoeuvre for a custom sweep-wing aircraft. Previous work showed that the use of domain randomisation to train with atmospheric disturbances improved the real-world performance of the controllers, leading to increased reward. This paper builds on the previous project, investigating enhancements and modifications to the learning process to further improve performance, and reduce final state error. These changes include modifying the observation by adding information about the airspeed to the standard aircraft state vector, employing further domain randomisation of the simulator, optimising the underlying RL algorithm and network structure, and changing to a continuous action space. Simulated investigations identified hyperparameter optimisation as achieving the most significant increase in reward performance. Several test cases were explored to identify the best combination of enhancements. Flight testing was performed, comparing a baseline model against some of the best performing test cases from simulation. Generally, test cases that performed better than the baseline in simulation also performed better in the real world. However, flight tests also identified limitations with the current numerical model. For some models, the chosen policy performs well in simulation yet stalls prematurely in reality, a problem known as the reality gap.
Publisher
Cambridge University Press (CUP)
Reference30 articles.
1. Reinforcement learning for UAV attitude control;Koch;ACM Trans. Cyber-Phys. Syst.,2019
2. [14] Moerland, T.M. , Broekens, J. , Plaat, A. and Jonker, C.M. Model-based Reinforcement Learning: A Survey, arXiv, 2020.
3. Human-level control through deep reinforcement learning;Mnih;Nature,2015
4. [1] Waldock, A. , Greatwood, C. , Salama, F. and Richardson, T. Learning to perform a perched landing on the ground using deep reinforcement learning, J. Intell. Rob. Syst. Theory Appl., 2018, 92, pp 685–704.
5. [9] Brockman, G. , Cheung, V. , Pettersson, L. , Schneider, J. , Schulman, J. , Tang, J. and Zaremba, W. OpenAI Gym, ArXiv, 2016.
Cited by
6 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献