Author:
Ben Sada Abdelkarim,Khelloufi Amar,Naouri Abdenacer,Ning Huansheng,Dhelim Sahraoui
Abstract
AbstractThe recent widespread of AI-powered real-time applications necessitates the use of edge computing for inference task offloading. Power constrained edge devices are required to balance between processing inference tasks locally or offload to edge servers. This decision is determined according to the time constraint demanded by the real-time nature of applications, and the energy constraint dictated by the device’s power budget. This problem is further exacerbated in the case of systems leveraging multiple local inference models varying in size and accuracy. In this work, we tackle the problem of assigning inference models to inference tasks either using local inference models or by offloading to edge servers under time and energy constraints while maximizing the overall accuracy of the system. This problem is shown to be strongly NP-hard and therefore, we propose a hybrid genetic algorithm (HGSTO) to solve this problem. We leverage the speed of simulated annealing (SA) with the accuracy of genetic algorithms (GA) to develop a hybrid, fast and accurate algorithm compared with classic GA, SA and Particle Swarm Optimization (PSO). Experiment results show that HGSTO achieved on-par or higher accuracy than GA while resulting in significantly lower scheduling times compared to other schemes.
Funder
National Natural Science Foundation of China
University College Dublin
Publisher
Springer Science and Business Media LLC
Cited by
1 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献