A High-Throughput Processor for GDN-Based Deep Learning Image Compression
-
Published:2023-05-18
Issue:10
Volume:12
Page:2289
-
ISSN:2079-9292
-
Container-title:Electronics
-
language:en
-
Short-container-title:Electronics
Author:
Shao Hu1ORCID, Liu Bingtao1, Li Zongpeng1, Yan Chenggang1, Sun Yaoqi1ORCID, Wang Tingyu1
Affiliation:
1. Institute of Information and Control, Hangzhou Dianzi University, Hangzhou 310000, China
Abstract
Deep learning-based image compression techniques can take advantage of the autoencoder’s benefits to achieve greater compression quality at the same bit rate as traditional image compression, which is more in line with user desires. Designing a high-performance processor that can increase the inference speed and efficiency of the deep learning image compression (DIC) network is important to make this technology more extensively employed in mobile devices. To the best of our knowledge, there is no dedicated processor that can accelerate DIC with low power consumption, and general-purpose network accelerators based on field programmable gate arrays (FPGA) cannot directly process compressed networks, so we propose a processor suitable for DIC in this paper. First, we analyze the image compression algorithm and quantize the data of the network into 16-bit fixed points using a dynamic hierarchical quantization. Then, we design an operation module, which is the core computational part for processing. It is composed of convolution, sampling, and normalization units, which pipeline the inference calculation for each layer of the network. To achieve high-throughput inference computing, the processing elements group (PEG) array with local buffers is developed for convolutional computation. Based on the common components in encoding and decoding, the sampling and normalization units are compatible with codec computation and utilized for image compression with time-sharing multiplexing. According to the control signal, the operation module could change the order of data flow through the three units so that they perform encoding and decoding operations, respectively. Based on these design methods and schemes, DIC is deployed into the Xilinx Zynq ZCU104 development board to achieve high-throughput image compression at 6 different bit rates. The experimental results show that the processor can run at 200 MHz and achieve 283.4 GOPS for the 16 bits fixed-point DIC network.
Subject
Electrical and Electronic Engineering,Computer Networks and Communications,Hardware and Architecture,Signal Processing,Control and Systems Engineering
Reference33 articles.
1. An end-to-end compression framework based on convolutional neural networks;Jiang;IEEE Trans. Circuits Syst. Video Technol.,2017 2. Image compression based on deep learning: A review;Yasin;Asian J. Res. Comput. Sci.,2021 3. The JPEG still picture compression standard;Wallace;IEEE Trans. Consum. Electron.,1992 4. Liu, Z., Liu, T., Wen, W., Jiang, L., Xu, J., Wang, Y., and Quan, G. (2018, January 24–29). DeepN-JPEG: A deep neural network favorable JPEG-based image compression framework. Proceedings of the 55th Annual Design Automation Conference, San Francisco, CA, USA. 5. Li, M., Zuo, W., Gu, S., Zhao, D., and Zhang, D. (2018, January 18–23). Learning convolutional networks for content-weighted image compression. Proceedings of the IEEE conference on computer Vision and Pattern Recognition, Salt Lake City, UT, USA.
Cited by
2 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
|
|