RepECN: Making ConvNets Better Again for Efficient Image Super-Resolution
Author:
Chen Qiangpu1ORCID, Qin Jinghui2, Wen Wushao1
Affiliation:
1. School of Computer Science and Engineering, Sun Yat-sen University, Guangzhou 510275, China 2. School of Information Engineering, Guangdong University of Technology, Guangzhou 510006, China
Abstract
Traditional Convolutional Neural Network (ConvNet, CNN)-based image super-resolution (SR) methods have lower computation costs, making them more friendly for real-world scenarios. However, they suffer from lower performance. On the contrary, Vision Transformer (ViT)-based SR methods have achieved impressive performance recently, but these methods often suffer from high computation costs and model storage overhead, making them hard to meet the requirements in practical application scenarios. In practical scenarios, an SR model should reconstruct an image with high quality and fast inference. To handle this issue, we propose a novel CNN-based Efficient Residual ConvNet enhanced with structural Re-parameterization (RepECN) for a better trade-off between performance and efficiency. A stage-to-block hierarchical architecture design paradigm inspired by ViT is utilized to keep the state-of-the-art performance, while the efficiency is ensured by abandoning the time-consuming Multi-Head Self-Attention (MHSA) and by re-designing the block-level modules based on CNN. Specifically, RepECN consists of three structural modules: a shallow feature extraction module, a deep feature extraction, and an image reconstruction module. The deep feature extraction module comprises multiple ConvNet Stages (CNS), each containing 6 Re-Parameterization ConvNet Blocks (RepCNB), a head layer, and a residual connection. The RepCNB utilizes larger kernel convolutions rather than MHSA to enhance the capability of learning long-range dependence. In the image reconstruction module, an upsampling module consisting of nearest-neighbor interpolation and pixel attention is deployed to reduce parameters and maintain reconstruction performance, while bicubic interpolation on another branch allows the backbone network to focus on learning high-frequency information. The extensive experimental results on multiple public benchmarks show that our RepECN can achieve 2.5∼5× faster inference than the state-of-the-art ViT-based SR model with better or competitive super-resolving performance, indicating that our RepECN can reconstruct high-quality images with fast inference.
Funder
National Natural Science Foundation of China GuangDong Basic and Applied Basic Research Foundation China Postdoctoral Science Foundation
Subject
Electrical and Electronic Engineering,Biochemistry,Instrumentation,Atomic and Molecular Physics, and Optics,Analytical Chemistry
Reference56 articles.
1. Dong, C., Loy, C.C., He, K., and Tang, X. (2014, January 6–12). Learning a Deep Convolutional Network for Image Super-Resolution. Proceedings of the European Conference on Computer Vision, Zurich, Switzerland. 2. Dong, C., Loy, C.C., and Tang, X. (2016, January 11–14). Accelerating the Super-Resolution Convolutional Neural Network. Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands. 3. Chen, H., Wang, Y., Guo, T., Xu, C., Deng, Y., Liu, Z., Ma, S., Xu, C., Xu, C., and Gao, W. (2021, January 20–25). Pre-Trained Image Processing Transformer. Proceedings of the Computer Vision and Pattern Recognition, Nashville, TN, USA. 4. Liang, J., Cao, J., Sun, G., Zhang, K., Van Gool, L., and Timofte, R. (2021, January 11–17). SwinIR: Image Restoration Using Swin Transformer. Proceedings of the International Conference on Computer Vision Workshops, Montreal, QC, Canada. 5. Ledig, C., Theis, L., Huszár, F., Caballero, J., Cunningham, A., Acosta, A., Aitken, A., Tejani, A., Totz, J., and Wang, Z. (2017, January 21–26). Photo-Realistic Single Image Super-Resolution Using a Generative Adversarial Network. Proceedings of the Computer Vision and Pattern Recognition, Honolulu, HI, USA.
|
|