Affiliation:
1. School of Electrical and Optoelectronic Engineering, West Anhui University, Lu’an 237012, China
Abstract
The input layer, hidden layer, and output layer are three models of the neural processors that make up feedforward neural networks (FNNs). Evolutionary algorithms have been extensively employed in training FNNs, which can correctly actualize any finite training sample set. In this paper, an enhanced marine predators algorithm (MPA) based on the ranking-based mutation operator (EMPA) was presented to train FNNs, and the objective was to attain the minimum classification, prediction, and approximation errors by modifying the connection weight and deviation value. The ranking-based mutation operator not only determines the best search agent and elevates the exploitation ability, but it also delays premature convergence and accelerates the optimization process. The EMPA integrates exploration and exploitation to mitigate search stagnation, and it has sufficient stability and flexibility to acquire the finest solution. To assess the significance and stability of the EMPA, a series of experiments on seventeen distinct datasets from the machine learning repository of the University of California Irvine (UCI) were utilized. The experimental results demonstrated that the EMPA has a quicker convergence speed, greater calculation accuracy, higher classification rate, strong stability and robustness, which is productive and reliable for training FNNs.
Funder
Start-up Fee for Scientific Research of High-level Talents in 2022
Subject
Process Chemistry and Technology,Chemical Engineering (miscellaneous),Bioengineering
Cited by
4 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献