Affiliation:
1. Department of Electrical and Computer Engineering, Democritus University of Thrace, Xanthi, Greece
2. Information Technologies Institute, Centre for Research and Technology Hellas (ITI-CERTH), Thessaloniki, Greece
Abstract
Distributed Machine learning has delivered considerable advances in training neural networks by leveraging parallel processing, scalability, and fault tolerance to accelerate the process and improve model performance. However, training of large-size models has exhibited numerous challenges, due to the gradient dependence that conventional approaches integrate. To improve the training efficiency of such models, gradient-free distributed methodologies have emerged fostering the gradient-independent parallel processing and efficient utilization of resources across multiple devices or nodes. However, such approaches, are usually restricted to specific applications, due to their conceptual limitations: computational and communicational requirements between partitions, limited partitioning solely into layers, limited sequential learning between the different layers, as well as training a potential model in solely synchronous mode. In this paper, we propose and evaluate, the Neuro-Distributed Cognitive Adaptive Optimization (ND-CAO) methodology, a novel gradient-free algorithm that enables the efficient distributed training of arbitrary types of neural networks, in both synchronous and asynchronous manner. Contrary to the majority of existing methodologies, ND-CAO is applicable to any possible splitting of a potential neural network, into blocks (partitions), with each of the blocks allowed to update its parameters fully asynchronously and independently of the rest of the blocks. Most importantly, no data exchange is required between the different blocks during training with the only information each block requires is the global performance of the model. Convergence of ND-CAO is mathematically established for generic neural network architectures, independently of the particular choices made, while four comprehensive experimental cases, considering different model architectures and image classification tasks, validate the algorithms’ robustness and effectiveness in both synchronous and asynchronous training modes. Moreover, by conducting a thorough comparison between synchronous and asynchronous ND-CAO training, the algorithm is identified as an efficient scheme to train neural networks in a novel gradient-independent, distributed, and asynchronous manner, delivering similar – or even improved results in Loss and Accuracy measures.
Subject
Artificial Intelligence,Computational Theory and Mathematics,Computer Science Applications,Theoretical Computer Science,Software
Reference95 articles.
1. A methodology using classification for traffic prediction: Featuring the impact of COVID-19;Liapis;Integrated Computer-Aided Engineering.,2021
2. Vulnerability prediction for secure healthcare supply chain service delivery;Islam;Integrated Computer-Aided Engineering.,2022
3. Dynamic learning rates for continual unsupervised learning;Fernández-Rodríguez;Integrated Computer-Aided Engineering.,2023
4. Fusion of multitemporal contextual information by neural networks for multisensor remote sensing image classification;Melgani;Integrated Computer-Aided Engineering.,2003
5. Krizhevsky A, Sutskever I, Hinton GE. Imagenet classification with deep convolutional neural networks. Advances in neural information processing systems. 2012; 25.