Advancements in On-Device Deep Neural Networks
-
Published:2023-08-21
Issue:8
Volume:14
Page:470
-
ISSN:2078-2489
-
Container-title:Information
-
language:en
-
Short-container-title:Information
Author:
Saravanan Kavya12, Kouzani Abbas Z.1ORCID
Affiliation:
1. School of Engineering, Deakin University, Geelong, VIC 3216, Australia 2. Department of Sensor and Biomedical Technology, Vellore Institute of Technology, Vellore 632014, India
Abstract
In recent years, rapid advancements in both hardware and software technologies have resulted in the ability to execute artificial intelligence (AI) algorithms on low-resource devices. The combination of high-speed, low-power electronic hardware and efficient AI algorithms is driving the emergence of on-device AI. Deep neural networks (DNNs) are highly effective AI algorithms used for identifying patterns in complex data. DNNs, however, contain many parameters and operations that make them computationally intensive to execute. Accordingly, DNNs are usually executed on high-resource backend processors. This causes an increase in data processing latency and energy expenditure. Therefore, modern strategies are being developed to facilitate the implementation of DNNs on devices with limited resources. This paper presents a detailed review of the current methods and structures that have been developed to deploy DNNs on devices with limited resources. Firstly, an overview of DNNs is presented. Next, the methods used to implement DNNs on resource-constrained devices are explained. Following this, the existing works reported in the literature on the execution of DNNs on low-resource devices are reviewed. The reviewed works are classified into three categories: software, hardware, and hardware/software co-design. Then, a discussion on the reviewed approaches is given, followed by a list of challenges and future prospects of on-device AI, together with its emerging applications.
Subject
Information Systems
Reference42 articles.
1. Fowers, J., Ovtcharov, K., Papamichael, M., Massengill, T., Liu, M., Lo, D., Alkalay, S., Haselman, M., Adams, L., and Ghandi, M. (2018, January 1–6). A configurable cloud-scale DNN processor for real-time AI. Proceedings of the 45th Annual International Symposium on Computer Architecture, Los Angeles, CA, USA. 2. Merenda, M., Porcaro, C., and Iero, D. (2020). Edge machine learning for ai-enabled IoT devices: A review. Sensors, 20. 3. Mishra, R., Gupta, H.P., and Dutta, T. (2020). A survey on deep neural network compression: Challenges, overview, and solutions. arXiv. 4. Implementation of DNNs on IoT devices;Zhichao;Neural Comput. Appl.,2020 5. Lane, N.D., Bhattacharya, S., Georgiev, P., Forlivesi, C., Jiao, L., Qendro, L., and Kawsar, F. (2016, January 11–14). DeepX: A Software Accelerator for Low-Power Deep Learning Inference on Mobile Devices. Proceedings of the 15th ACM/IEEE International Conference on Information Processing in Sensor Networks (IPSN), Vienna, Austria.
Cited by
3 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
|
|