A Parametrizable High-Level Synthesis Library for Accelerating Neural Networks on FPGAs
-
Published:2021-03-15
Issue:5
Volume:93
Page:513-529
-
ISSN:1939-8018
-
Container-title:Journal of Signal Processing Systems
-
language:en
-
Short-container-title:J Sign Process Syst
Author:
Kalms LesterORCID, Rad Pedram AminiORCID, Ali MuhammadORCID, Iskander Arsany, Göhringer DianaORCID
Abstract
AbstractIn recent years, Convolutional Neural Network CNN have been incorporated in a large number of applications, including multimedia retrieval and image classification. However, CNN based algorithms are computationally and resource intensive and therefore difficult to be used in embedded systems. FPGA based accelerators are becoming more and more popular in research and industry due to their flexibility and energy efficiency. However, the available resources and the size of the on-chip memory can limit the performance of the FPGA accelerator for CNN. This work proposes an High-Level Synthesis HLS library for CNN algorithms. It contains seven different streaming-capable CNN (plus two conversion) functions for creating large neural networks with deep pipelines. The different functions have many parameter settings (e.g. for resolution, feature maps, data types, kernel size, parallelilization, accuracy, etc.), which also enable compile-time optimizations. Our functions are integrated into the HiFlipVX library, which is an open source HLS FPGA library for image processing and object detection. This offers the possibility to implement different types of computer vision applications with one library. Due to the various configuration and parallelization possibilities of the library functions, it is possible to implement a high-performance, scalable and resource-efficient system, as our evaluation of the MobileNets algorithm shows.
Publisher
Springer Science and Business Media LLC
Subject
Hardware and Architecture,Modeling and Simulation,Information Systems,Signal Processing,Theoretical Computer Science,Control and Systems Engineering
Reference39 articles.
1. Abadi, M., Barham, P., Chen, J., Chen, Z., Davis, A., Dean, J., Devin, M., Ghemawat, S., Irving, G., Isard, M., et al. (2016). Tensorflow: A system for large-scale machine learning. In 12th USENIX symposium on operating systems design and implementation (OSDI 16). https://github.com/tensorflow/models/blob/master/research/slim/nets/mobilenet_v1.md (pp. 265–283). 2. Akgün, G., Kalms, L., Göhringer, D. (2020). Resource efficient dynamic voltage and frequency scaling on xilinx fpgas. In International symposium on applied reconfigurable computing (ARC) (pp. 178–192). 3. Chen, Y., He, J., Zhang, X., Hao, C., Chen, D. (2019). Cloud-dnn: an open framework for mapping dnn models to cloud fpgas. In Proceedings of the international symposium on field-programmable gate arrays (FPGA) (pp. 73–82). https://doi.org/10.1145/3289602.3293915. 4. Chen, Y., Luo, T., Liu, S., Zhang, S., He, L., Wang, J., Li, L., Chen, T., Xu, Z., Sun, N., Temam, O. (2014). Dadiannao: a machine-learning supercomputer. In 47th annual IEEE/ACM international symposium on microarchitecture (pp. 609–622). 5. Giduthuri, R., & Pulli, K. (2016). Openvx: A framework for accelerating computer vision. In SIGGRAPH ASIA 2016 Courses (pp. 14:1–14:50). https://doi.org/10.1145/2988458.2988513.
Cited by
11 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
|
|