A Lightweight Recurrent Grouping Attention Network for Video Super-Resolution
Author:
Zhu Yonggui1ORCID, Li Guofang2ORCID
Affiliation:
1. School of Data Science and Intelligent Media, Communication University of China, Beijing 100024, China 2. School of Information and Communication Engineering, Communication University of China, Beijing 100024, China
Abstract
Effective aggregation of temporal information of consecutive frames is the core of achieving video super-resolution. Many scholars have utilized structures such as sliding windows and recurrences to gather the spatio-temporal information of frames. However, although the performances of constructed video super-resolution models are improving, the sizes of the models are also increasing, exacerbating the demand on the equipment. Thus, to reduce the stress on the device, we propose a novel lightweight recurrent grouping attention network. The parameters of this model are only 0.878 M, which is much lower than the current mainstream model for studying video super-resolution. We have designed a forward feature extraction module and a backward feature extraction module to collect temporal information between consecutive frames from two directions. Moreover, a new grouping mechanism is proposed to efficiently collect spatio-temporal information of the reference frame and its neighboring frames. The attention supplementation module is presented to further enhance the information gathering range of the model. The feature reconstruction module aims to aggregate information from different directions to reconstruct high-resolution features. Experiments demonstrate that our model achieves state-of-the-art performance on multiple datasets.
Funder
National Natural Science Foundation of China Fundamental Research Funds for the Central Universities
Subject
Electrical and Electronic Engineering,Biochemistry,Instrumentation,Atomic and Molecular Physics, and Optics,Analytical Chemistry
Reference56 articles.
1. Cross-Frame Transformer-Based Spatio-Temporal Video Super-Resolution;Zhang;IEEE Trans. Broadcast.,2022 2. Pan, J., Bai, H., Dong, J., Zhang, J., and Tang, J. (2021, January 10–17). Deep Blind Video Super-resolution. Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada. 3. Xiao, J., Jiang, X., Zheng, N., Yang, H., Yang, Y., Yang, Y., Li, D., and Lam, K. (2022). Online Video Super-Resolution with Convolutional Kernel Bypass Graft. IEEE Trans. Multimed., 1–16. 4. Wang, Y., Isobe, T., Jia, X., Tao, X., Lu, H., and Tai, Y. (2023, January 14–19). Compression-Aware Video Super-Resolution. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA. 5. Learning a Local-Global Alignment Network for Satellite Video Super-Resolution;Jin;IEEE Geosci. Remote Sens. Lett.,2023
Cited by
1 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
1. Bidirectional recurrent deformable alignment network for video super-resolution;International Conference on Image, Signal Processing, and Pattern Recognition (ISPP 2024);2024-06-13
|
|