Abstract
Active flow control for drag reduction with reinforcement learning (RL) is performed in the wake of a two-dimensional square bluff body at laminar regimes with vortex shedding. Controllers parametrised by neural networks are trained to drive two blowing and suction jets that manipulate the unsteady flow. The RL with full observability (sensors in the wake) discovers successfully a control policy that reduces the drag by suppressing the vortex shedding in the wake. However, a non-negligible performance degradation (
$\sim$
50 % less drag reduction) is observed when the controller is trained with partial measurements (sensors on the body). To mitigate this effect, we propose an energy-efficient, dynamic, maximum entropy RL control scheme. First, an energy-efficiency-based reward function is proposed to optimise the energy consumption of the controller while maximising drag reduction. Second, the controller is trained with an augmented state consisting of both current and past measurements and actions, which can be formulated as a nonlinear autoregressive exogenous model, to alleviate the partial observability problem. Third, maximum entropy RL algorithms (soft actor critic and truncated quantile critics) that promote exploration and exploitation in a sample-efficient way are used, and discover near-optimal policies in the challenging case of partial measurements. Stabilisation of the vortex shedding is achieved in the near wake using only surface pressure measurements on the rear of the body, resulting in drag reduction similar to that in the case with wake sensors. The proposed approach opens new avenues for dynamic flow control using partial measurements for realistic configurations.
Funder
Engineering and Physical Sciences Research Council
Publisher
Cambridge University Press (CUP)
Reference80 articles.
1. Linear feedback stabilization of laminar vortex shedding based on a point vortex model;Protas;Phys. Fluids,2004
2. Reinforcement learning of control strategies for reducing skin friction drag in a fully developed turbulent channel flow;Sonoda;J. Fluid Mech.,2023
3. Dielectric barrier discharge plasma actuators for flow control;Corke;Annu. Rev. Fluid Mech.,2010
4. Sutton, R.S. & Barto, A.G. 2018 Reinforcement Learning: An Introduction. MIT Press.
5. Ziebart, B.D. , Maas, A. , Bagnell, J.A. & Dey, A.K. 2008 Maximum entropy inverse reinforcement learning. In Association for the Advancement of Artificial Intelligence, pp. 1433–1438. AAAI.
Cited by
5 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献