Abstract
AbstractNowadays, artificial intelligence (AI) has become a crucial Key Enabling Technology with extensive application in diverse industrial sectors. Recently, considerable focus has been directed towards utilizing AI for the development of optimal control in industrial processes. In particular, reinforcement learning (RL) techniques have made significant advancements, enabling their application to data-driven problem-solving for the control of complex systems. Since industrial manufacturing processes can be treated as MIMO non-linear systems, RL can be used to develop complex data-driven intelligent decision-making or control systems. In this work, the workflow for developing a RL application for industrial manufacturing processes, including reward function setup, development of reduced order models and control policy construction, is addressed, and a new process-based reward function is proposed. To showcase the proposed approach, a case study is developed with reference to a wire arc additive manufacturing (WAAM) process. Based on experimental tests, a Reduced Order Model of the system is obtained and a Deep Deterministic Policy Gradient Controller is trained with aim to produce a simple geometry. Particular attention is given to the sim-to-real process by developing a WAAM simulator which allows to simulate the process in a realistic environment and to generate the code to be deployed on the motion platform controller.
Funder
Università degli Studi di Napoli Federico II
Publisher
Springer Science and Business Media LLC
Subject
Artificial Intelligence,Industrial and Manufacturing Engineering,Software
Reference59 articles.
1. Athans, M. (1971). The role and use of the stochastic linear-quadratic-Gaussian problem in control system design. IEEE Transactions on Automatic Control, 16(6), 529–552.
2. Bellman, R. (1966). Dynamic programming. Science, 153(3731), 34–37.
3. Caggiano, A., Giulio, M., & Luigi, N. (2023). Smart tool wear monitoring of CFRP/CFRP stack drilling using autoencoders and memory-based neural networks. Applied Sciences (switzerland). https://doi.org/10.3390/app13053307
4. Dann, C., Mansour, Y., Mohri, M., Sekhari, A., & Sridharan, K. (2022). Guarantees for epsilon-greedy reinforcement learning with function approximation. International conference on machine learning (pp. 4666–4689). PMLR.
5. Datta, L. (2020). A survey on activation functions and their relation with Xavier and He normal initialization. Preprint retrieved from https://arxiv.org/abs/2004.06632