Abstract
Due to various breakthroughs and advancements in machine learning and computer architectures, machine learning models are beginning to proliferate through embedded platforms. Some of these machine learning models cover a range of applications including computer vision, speech recognition, healthcare efficiency, industrial IoT, robotics and many more. However, there is a critical limitation in implementing ML algorithms efficiently on embedded platforms: the computational and memory expense of many machine learning models can make them unsuitable in resource-constrained environments. Therefore, to efficiently implement these memory-intensive and computationally expensive algorithms in an embedded computing environment, innovative resource management techniques are required at the hardware, software and system levels. To this end, we present a novel quality-of-service based resource allocation scheme that uses feedback control to adjust compute resources dynamically to cope with the varying and unpredictable workloads of ML applications while still maintaining an acceptable level of service to the user. To evaluate the feasibility of our approach we implemented a feedback control scheduling simulator that was used to analyze our framework under various simulated workloads. We also implemented our framework as a Linux kernel module running on a virtual machine as well as a Raspberry Pi 4 single board computer. Results illustrate that our approach was able to maintain a sufficient level of service without overloading the processor as well as providing an energy savings of almost 20% as compared to the native resource management in Linux.
Subject
Electrical and Electronic Engineering,Computer Networks and Communications,Hardware and Architecture,Signal Processing,Control and Systems Engineering
Reference40 articles.
1. Going deeper with convolutions;Szegedy;Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR),2015
2. Deep residual learning for image recognition;He;Proceedings of the IEEE Computer Society Conference Computer Vision Pattern Recognition,2016
3. Large-scale evolution of image classifiers;Real;Proceedings of the 34th International Conference Machine Learning ICML,2017
4. EfficientNet: Rethinking model scaling for convolutional neural networks;Tan;Proceedings of the 36th International Conference Machine Learning ICML 2019,2019
5. Deep Neural Networks for Acoustic Modeling in Speech Recognition: The Shared Views of Four Research Groups
Cited by
1 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献