Affiliation:
1. Artificial Intelligence Computing Research Laboratory Electronics and Telecommunications Research Institute Daejeon Republic of Korea
2. University of Science and Technology Daejeon Republic of Korea
Abstract
AbstractThis paper introduces a framework for optimizing deep‐learning models on microcontrollers (MCUs) that is crucial in today's expanding embedded device market. We focus on model optimization techniques, particularly pruning and quantization, to enhance the performance of neural networks within the limited resources of MCUs. Our approach combines automatic iterative optimization and code generation, simplifying MCU model deployment without requiring extensive hardware knowledge. Based on experiments with architectures, such as ResNet‐8 and MobileNet v2, our framework substantially reduces the model size and enhances inference speed that are crucial for MCU efficiency. Compared with TensorFlow Lite for MCUs, our optimizations for MobileNet v2 reduce static random‐access memory use by 51%–57% and flash use by 17%–62%, while increasing inference speed by approximately 1.55 times. These advancements highlight the impact of our method on performance and memory efficiency, demonstrating its value in embedded artificial intelligence and broad applicability in MCU‐based neural network optimization.
Funder
Defense Acquisition Program Administration
Reference36 articles.
1. EMOS: Enhanced moving object detection and classification via sensor fusion and noise filtering
2. A robust collision prediction and detection method based on neural network for autonomous delivery robots
3. Deep learning on microcontrollers
4. MCUNet: Tiny deep learning on IoT devices;Lin J.;Adv. Neural Inf. Process. Syst.,2020
5. Micronets: Neural network architectures for deploying tinyml applications on commodity microcontrollers;Banbury C.;Proc. Mach. Learn. Syst.,2021