Abstract
Convolutional Neural Networks (ConvNets) can be shrunk to fit embedded CPUs adopted on mobile end-nodes, like smartphones or drones. The deployment onto such devices encompasses several algorithmic level optimizations, e.g., topology restructuring, pruning, and quantization, that reduce the complexity of the network, ensuring less resource usage and hence higher speed. Several studies revealed remarkable performance, paving the way towards real-time inference on low power cores. However, continuous execution at maximum speed is quite unrealistic due to a fast increase of the on-chip temperature. Indeed, proper thermal management is paramount to guarantee silicon reliability and a safe user experience. Power management schemes, like voltage lowering and frequency scaling, are common knobs to control the thermal stability. Obviously, this implies a performance degradation, often not considered during the training and optimization stages. The objective of this work is to present the performance assessment of embedded ConvNets under thermal management. Our study covers the behavior of two control policies, namely reactive and proactive, implemented through the Dynamic Voltage-Frequency Scaling (DVFS) mechanism available on commercial embedded CPUs. As benchmarks, we used four state-of-the-art ConvNets for computer vision flashed into the ARM Cortex-A15 CPU. With the collected results, we aim to show the existing temperature-performance trade-off and give a more realistic analysis of the maximum performance achievable. Moreover, we empirically demonstrate the strict relationship between the on-chip thermal behavior and the hyper-parameters of the ConvNet, revealing optimization margins for a thermal-aware design of neural network layers.
Subject
Electrical and Electronic Engineering,Computer Networks and Communications,Hardware and Architecture,Signal Processing,Control and Systems Engineering
Cited by
9 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
1. MOC: Multi-Objective Mobile CPU-GPU Co-Optimization for Power-Efficient DNN Inference;2023 IEEE/ACM International Conference on Computer Aided Design (ICCAD);2023-10-28
2. Automating CPU Dynamic Thermal Control for High Performance Computing;2022 22nd IEEE International Symposium on Cluster, Cloud and Internet Computing (CCGrid);2022-05
3. On the Efficiency of AdapTTA: An Adaptive Test-Time Augmentation Strategy for Reliable Embedded ConvNets;VLSI-SoC: Technology Advancement on SoC Design;2022
4. AdapTTA: Adaptive Test-Time Augmentation for Reliable Embedded ConvNets;2021 IFIP/IEEE 29th International Conference on Very Large Scale Integration (VLSI-SoC);2021-10-04
5. TVFS: Topology Voltage Frequency Scaling for Reliable Embedded ConvNets;IEEE Transactions on Circuits and Systems II: Express Briefs;2021-02