Affiliation:
1. Syracuse University, Syracuse, NY, USA
2. University of Southern California, Los Angeles, CA, USA
3. City University of New York, New York, NY, USA
Abstract
With the recent advance of wearable devices and Internet of Things (IoTs), it becomes attractive to implement the Deep Convolutional Neural Networks (DCNNs) in embedded and portable systems. Currently, executing the software-based DCNNs requires high-performance servers, restricting the widespread deployment on embedded and mobile IoT devices. To overcome this obstacle, considerable research efforts have been made to develop highly-parallel and specialized DCNN accelerators using GPGPUs, FPGAs or ASICs.
Stochastic Computing (SC), which uses a bit-stream to represent a number within [-1, 1] by counting the number of ones in the bit-stream, has high potential for implementing DCNNs with high scalability and ultra-low hardware footprint. Since multiplications and additions can be calculated using AND gates and multiplexers in SC, significant reductions in power (energy) and hardware footprint can be achieved compared to the conventional binary arithmetic implementations. The tremendous savings in power (energy) and hardware resources allow immense design space for enhancing scalability and robustness for hardware DCNNs.
This paper presents SC-DCNN, the first comprehensive design and optimization framework of SC-based DCNNs, using a bottom-up approach. We first present the designs of function blocks that perform the basic operations in DCNN, including inner product, pooling, and activation function. Then we propose four designs of feature extraction blocks, which are in charge of extracting features from input feature maps, by connecting different basic function blocks with joint optimization. Moreover, the efficient weight storage methods are proposed to reduce the area and power (energy) consumption. Putting all together, with feature extraction blocks carefully selected, SC-DCNN is holistically optimized to minimize area and power (energy) consumption while maintaining high network accuracy. Experimental results demonstrate that the LeNet5 implemented in SC-DCNN consumes only 17
mm
2
area and 1.53 W power, achieves throughput of 781250 images/s, area efficiency of 45946 images/s/
mm
2
, and energy efficiency of 510734 images/J.
Funder
the seedling fund of DARPA SAGA program
Publisher
Association for Computing Machinery (ACM)
Subject
Computer Graphics and Computer-Aided Design,Software
Reference46 articles.
1. Stanford cs class cs231n: Convolutional neural networks for visual recognition 2016. URL http://cs231n.github. io/convolutional-networks/. Stanford cs class cs231n: Convolutional neural networks for visual recognition 2016. URL http://cs231n.github. io/convolutional-networks/.
2. Convolutional neural networks (lenet) 2016. URL http://deeplearning.net/tutorial/lenet.html# motivation. Convolutional neural networks (lenet) 2016. URL http://deeplearning.net/tutorial/lenet.html# motivation.
3. Nangate 45nm Open Library Nangate Inc. 2009. URL http: //www.nangate.com/ . Nangate 45nm Open Library Nangate Inc. 2009. URL http: //www.nangate.com/ .
4. TrueNorth: Design and Tool Flow of a 65 mW 1 Million Neuron Programmable Neurosynaptic Chip
5. R. Andri L. Cavigelli D. Rossi and L. Benini. Yodann: An ultra-low power convolutional neural network accelerator based on binary weights. arXiv preprint arXiv:1606.05487 2016. R. Andri L. Cavigelli D. Rossi and L. Benini. Yodann: An ultra-low power convolutional neural network accelerator based on binary weights. arXiv preprint arXiv:1606.05487 2016.
Cited by
69 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献