Author:
Song Zhourui,Liu Zhenyu,Wang Dongsheng
Abstract
The heavy burdens of computation and off-chip traffic impede deploying the large scale convolution neural network on embedded platforms. As CNN is attributed to the strong endurance to computation errors, employing block floating point (BFP) arithmetics in CNN accelerators could save the hardware cost and data traffics efficiently, while maintaining the classification accuracy. In this paper, we verify the effects of word width definitions in BFP to the CNN performance without retraining. Several typical CNN models, including VGG16, ResNet-18, ResNet-50 and GoogLeNet, were tested in this paper. Experiments revealed that 8-bit mantissa, including sign bit, in BFP representation merely induced less than 0.3% accuracy loss. In addition, we investigate the computational errors in theory and develop the noise-to-signal ratio (NSR) upper bound, which provides the promising guidance for BFP based CNN engine design.
Publisher
Association for the Advancement of Artificial Intelligence (AAAI)
Cited by
9 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
1. Mirage: An RNS-Based Photonic Accelerator for DNN Training;2024 ACM/IEEE 51st Annual International Symposium on Computer Architecture (ISCA);2024-06-29
2. Design of an Efficient Deep Neural Network Accelerator Based on Block Posit Number Representation;2024 International VLSI Symposium on Technology, Systems and Applications (VLSI TSA);2024-04-22
3. FIGNA: Integer Unit-Based Accelerator Design for FP-INT GEMM Preserving Numerical Accuracy;2024 IEEE International Symposium on High-Performance Computer Architecture (HPCA);2024-03-02
4. Improving finite-difference time-domain memory bandwidth by using block floating-point arithmetic;2023 31st Telecommunications Forum (TELFOR);2023-11-21
5. Machine Learning;Design for Embedded Image Processing on FPGAs;2023-09-05