Affiliation:
1. Government College University
2. Jiangsu University
Abstract
Activation functions play a crucial role in enabling neural networks to carry out tasks with increased flexibility by introducing non-linearity. The selection of appropriate activation functions becomes even more crucial, especially in the context of deeper networks where the objective is to learn more intricate patterns. Among various deep learning tools, Convolutional Neural Networks (CNNs) stand out for their exceptional ability to learn complex visual patterns. In practice, ReLu is commonly employed in convolutional layers of CNNs, yet other activation functions like Swish can demonstrate superior training performance while maintaining good testing accuracy on different datasets. This paper presents an optimally refined strategy for deep learning-based image classification tasks by incorporating CNNs with advanced activation functions and an adjustable setting of layers. A thorough analysis has been conducted to support the effectiveness of various activation functions when coupled with the favorable softmax loss, rendering them suitable for ensuring a stable training process. The results obtained on the CIFAR-10 dataset demonstrate the favorability and stability of the adopted strategy throughout the training process.
Reference45 articles.
1. Alpaydin, E. (2020). Introduction to Machine Learning.
MIT press.
2. Bonnell, J. A. (2011). Implementation of a New
Sigmoid Function in Back propagation Neural Networks
(Master thesis, East Tennessee State University).
3. Chaithanya, B. N., Swasthika Jain, T. J., Usha Ruby, A.,
& Parveen, A. (2021). An approach to categorize chest Xray
images using sparse categorical cross entropy.
Indonesian Journal of Electrical Engineering and
Computer Science, 24(3), 1700-1710.
4. Covington, P., Adams, J., & Sargin, E. (2016,
September). Deep neural networks for youtube
recommendations. In Proceedings of the 10th ACM
Conference on Recommender Systems, (pp. 191-198).