Author:
Reyad Mohamed,Sarhan Amany M.,Arafa M.
Abstract
AbstractDeep Neural Networks (DNNs) are widely regarded as the most effective learning tool for dealing with large datasets, and they have been successfully used in thousands of applications in a variety of fields. Based on these large datasets, they are trained to learn the relationships between various variables. The adaptive moment estimation (Adam) algorithm, a highly efficient adaptive optimization algorithm, is widely used as a learning algorithm in various fields for training DNN models. However, it needs to improve its generalization performance, especially when training with large-scale datasets. Therefore, in this paper, we propose HN Adam, a modified version of the Adam Algorithm, to improve its accuracy and convergence speed. The HN_Adam algorithm is modified by automatically adjusting the step size of the parameter updates over the training epochs. This automatic adjustment is based on the norm value of the parameter update formula according to the gradient values obtained during the training epochs. Furthermore, a hybrid mechanism was created by combining the standard Adam algorithm and the AMSGrad algorithm. As a result of these changes, the HN_Adam algorithm, like the stochastic gradient descent (SGD) algorithm, has good generalization performance and achieves fast convergence like other adaptive algorithms. To test the proposed HN_Adam algorithm performance, it is evaluated to train a deep convolutional neural network (CNN) model that classifies images using two different standard datasets: MNIST and CIFAR-10. The algorithm results are compared to the basic Adam algorithm and the SGD algorithm, in addition to other five recent SGD adaptive algorithms. In most comparisons, the HN Adam algorithm outperforms the compared algorithms in terms of accuracy and convergence speed. AdaBelief is the most competitive of the compared algorithms. In terms of testing accuracy and convergence speed (represented by the consumed training time), the HN-Adam algorithm outperforms the AdaBelief algorithm by an improvement of 1.0% and 0.29% for the MNIST dataset, and 0.93% and 1.68% for the CIFAR-10 dataset, respectively.
Publisher
Springer Science and Business Media LLC
Subject
Artificial Intelligence,Software
Reference66 articles.
1. Alzubaidi L, Zhang J, Humaidi AJ (2021) Review of deep learning: concepts, CNN architectures, challenges, applications, future directions. J Big Data 8:53
2. Michael G, Kaldewey T, Tam D (2017) Optimizing the efficiency of deep learning through accelerator virtualization. IBM J Res Dev 61:121–1211. https://doi.org/10.1147/JRD.2017.2716598
3. Maurizio C, Beatrice B, Alberto M, Muhammad S, Guido M (2020) An updated survey of efficient hardware architectures for accelerating deep convolutional neural networks. J Fut Inter 12:113
4. Pouyanfar S, Sadiq S, Yan Y (2018) A survey on deep learning: algorithms, techniques, and applications. ACM Comput Surv 51:5
5. Hassen L, Slim B, Ali L, Chih CH, Lamjed BS (2021) Deep convolutional neural network architecture design as a bi-level optimization problem. J Neuro Comput 439:44–62
Cited by
73 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献