An Improved BGE-Adam Optimization Algorithm Based on Entropy Weighting and Adaptive Gradient Strategy
Author:
Shao Yichuan1, Wang Jiantao2ORCID, Sun Haijing1ORCID, Yu Hao2ORCID, Xing Lei3, Zhao Qian4, Zhang Le1ORCID
Affiliation:
1. School of Intelligent Science Engineering, Shenyang University, Shenyang 110044, China 2. School of Information Engineering, Shenyang University, Shenyang 110044, China 3. School of Chemistry and Chemical Engineering, University of Surrey, Guildford GU2 7XH, UK 4. School of Science, Shenyang University of Technology, Shenyang 110044, China
Abstract
This paper introduces an enhanced variant of the Adam optimizer—the BGE-Adam optimization algorithm—that integrates three innovative technologies to augment the adaptability, convergence, and robustness of the original algorithm under various training conditions. Firstly, the BGE-Adam algorithm incorporates a dynamic β parameter adjustment mechanism that utilizes the rate of gradient variations to dynamically adjust the exponential decay rates of the first and second moment estimates (β1 and β2), the adjustment of β1 and β2 is symmetrical, which means that the rules that the algorithm considers when adjusting β1 and β2 are the same. This design helps to maintain the consistency and balance of the algorithm, allowing the optimization algorithm to adaptively capture the trending movements of gradients. Secondly, it estimates the direction of future gradients by a simple gradient prediction model, combining historic gradient information with the current gradient. Lastly, entropy weighting is integrated into the gradient update step. This strategy enhances the model’s exploratory nature by introducing a certain amount of noise, thereby improving its adaptability to complex loss surfaces. Experimental results on classical datasets, MNIST and CIFAR10, and gastrointestinal disease medical datasets demonstrate that the BGE-Adam algorithm has improved convergence and generalization capabilities. In particular, on the specific medical image gastrointestinal disease test dataset, the BGE-Adam optimization algorithm achieved an accuracy of 69.36%, a significant improvement over the 67.66% accuracy attained using the standard Adam algorithm; on the CIFAR10 test dataset, the accuracy of the BGE-Adam algorithm reached 71.4%, which is higher than the 70.65% accuracy of the Adam optimization algorithm; and on the MNIST dataset, the BGE-Adam algorithm’s accuracy was 99.34%, surpassing the Adam optimization algorithm’s accuracy of 99.23%. The BGE-Adam optimization algorithm exhibits better convergence and robustness. This research not only demonstrates the effectiveness of the combination of these three technologies but also provides new perspectives for the future development of deep learning optimization algorithms.
Funder
Liaoning Provincial Department of Education’s Higher Education Foundation Research Project (General Project), Shenyang University of Technology Liaoning Provincial Department of Education Science “14th Five-Year Plan” Ministry of Education’s “Chunhui Plan” Liaoning Provincial Department of Education’s Basic Research Project “Training and Application of Vertical Field Multi-Mode Deep Neural Network Model” Shenyang Science and Technology Plan “Special Mission for Leech Breeding and Traditional Chinese Medicine Planting in Dengshibao Town, Faku County”
Reference25 articles.
1. Anjum, M., and Shahab, S. (2023). Improving Autonomous Vehicle Controls and Quality Using Natural Language Processing-Based Input Recognition Model. Sustainability, 15. 2. Kingma, D.P., and Ba, J. (2014). Adam: A method for stochastic optimization. arXiv. 3. Kashyap, R. (2023). A survey of deep learning optimizers–First and second order methods. arXiv. 4. Zhang, Z., Ma, L., Li, Z., and Wu, C. (2018). Normalized Direction-preserving Adam. arXiv. 5. A modified Adam algorithm for deep neural network optimization;Reyad;Neural Comput. Appl.,2023
Cited by
1 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
|
|