Affiliation:
1. School of Information and Communication Engineering, Communication University of China, Beijing 100024, China
Abstract
Ultra-high-definition (UHD) video has brought new challenges to objective video quality assessment (VQA) due to its high resolution and high frame rate. Most existing VQA methods are designed for non-UHD videos—when they are employed to deal with UHD videos, the processing speed will be slow and the global spatial features cannot be fully extracted. In addition, these VQA methods usually segment the video into multiple segments, predict the quality score of each segment, and then average the quality score of each segment to obtain the quality score of the whole video. This breaks the temporal correlation of the video sequences and is inconsistent with the characteristics of human visual perception. In this paper, we present a no-reference VQA method, aiming to effectively and efficiently predict quality scores for UHD videos. First, we construct a spatial distortion feature network based on a super-resolution model (SR-SDFNet), which can quickly extract the global spatial distortion features of UHD videos. Then, to aggregate the spatial distortion features of each UHD frame, we propose a time fusion network based on a reinforcement learning model (RL-TFNet), in which the actor network continuously combines multiple frame features extracted by SR-SDFNet and outputs an action to adjust the current quality score to approximate the subjective score, and the critic network outputs action values to optimize the quality perception of the actor network. Finally, we conduct large-scale experiments on UHD VQA databases and the results reveal that, compared to other state-of-the-art VQA methods, our method achieves competitive quality prediction performance with a shorter runtime and fewer model parameters.
Funder
National Key R&D Program of China
Subject
Electrical and Electronic Engineering,Biochemistry,Instrumentation,Atomic and Molecular Physics, and Optics,Analytical Chemistry
Reference45 articles.
1. St-greed: Space-time generalized entropic differences for frame rate dependent video quality prediction;Madhusudana;IEEE Trans. Image Process.,2021
2. Quality assessment for video with degradation along salient trajectories;Wu;IEEE Trans. Multimed.,2019
3. Study of subjective quality and objective blind quality prediction of stereoscopic videos;Appina;IEEE Trans. Image Process.,2019
4. Korhonen, J., Su, Y., and You, J. (2020, January 12–16). Blind natural video quality prediction via statistical temporal features and deep spatial features. Proceedings of the 28th ACM International Conference on Multimedia, Seattle, WA, USA.
5. Multi-model standard for bitstream-, pixel-based and hybrid video quality assessment of uhd/4k: Itu-t p.1204;Raake;IEEE Access,2020
Cited by
4 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献