Affiliation:
1. Inria Chile Research Center, Av. Apoquindo 2827, Chile
Abstract
Physics-informed neural networks formulation allows the neural network to be trained by both the training data and prior domain knowledge about the physical system that models the data. In particular, it has a loss function for the data and the physics, where the latter is the deviation from a partial differential equation describing the system. Conventionally, both loss functions are combined by a weighted sum, whose weights are usually chosen manually. It is known that balancing between different loss terms can make the training process more efficient. In addition, it is necessary to find the optimal architecture of the neural network in order to find a hypothesis set in which is easier to train the PINN. In our work, we propose a multi-objective optimization approach to find the optimal value for the loss function weighting, as well as the optimal activation function, number of layers, and number of neurons for each layer. We validate our results on the Poisson, Burgers, and advection-diffusion equations and show that we are able to find accurate approximations of the solutions using optimal hyperparameters.
Reference18 articles.
1. G. Allaire, Numerical Analysis and Optimization: An Introduction to Mathematical Modelling and Numerical Simulation, Oxford university press, Oxford, UK, 2007.
2. Spectral and finite difference solutions of the Burgers equation;Basdevant;Computers & fluids,1986
3. J. Branke et al. (eds), Multiobjective Optimization, Lecture Notes in Computer Science, Vol. 5252, Springer–Verlag, Berlin/Heidelberg, 2008.
4. The multi-task learning with an application of Pareto improvement
5. Z. Chen, V. Badrinarayanan, C. Lee and A. Rabinovich, Gradnorm: Gradient normalization for adaptive loss balancing in deep multitask networks, in: ICML, Proceedings of Machine Learning Research, 2018, pp. 794–803.