Affiliation:
1. Department of Electrical and Computer Engineering, University of Patras, Rio, 26504 Patras, Greece
2. School of Mechanical Engineering, National Technical University of Athens, 15772 Athens, Greece
Abstract
The goal of this paper is to find a stabilizing and optimal control policy for a class of systems dictated by Euler–Lagrange dynamics, that also satisfies predetermined response criteria. The proposed methodology builds upon two stages. Initially, a neural network is trained online via an iterative process to capture the system dynamics, which are assumed to be unknown. Subsequently, a successive approximation algorithm is applied, employing the acquired dynamics from the previous step, to find a near-optimal control law that takes into consideration prescribed performance specifications, such as convergence speed and steady-state error. In addition, we concurrently guarantee that the system evolves exclusively within the compact set for which sufficient approximation capabilities have been acquired. Finally, we validate our claims through various simulated studies that confirm the success of both the identification process and the minimization of the cost function.
Funder
Hellenic Foundation for Research and Innovation
Subject
Fluid Flow and Transfer Processes,Computer Science Applications,Process Chemistry and Technology,General Engineering,Instrumentation,General Materials Science
Reference30 articles.
1. Powell, W.B., and Ryzhov, I.O. (2012). Reinforcement Learning and Approximate Dynamic Programming for Feedback Control, John Wiley & Sons, Ltd.. Chapter 18.
2. Kirk, D. (2004). Optimal Control Theory: An Introduction, Dover Publications.
3. Reinforcement learning and adaptive dynamic programming for feedback control;Lewis;IEEE Circuits Syst. Mag.,2009
4. Elements of intelligence;Werbos;Cybernetica,1968
5. Discrete-Time Nonlinear HJB Solution Using Approximate Dynamic Programming: Convergence Proof;Lewis;IEEE Trans. Syst. Man Cybern. Part B Cybern.,2008