Author:
Ke Junjie,Zhang Tianhao,Wang Yilin,Milanfar Peyman,Yang Feng
Abstract
No-reference video quality assessment (NR-VQA) for user generated content (UGC) is crucial for understanding and improving visual experience. Unlike video recognition tasks, VQA tasks are sensitive to changes in input resolution. Since large amounts of UGC videos nowadays are 720p or above, the fixed and relatively small input used in conventional NR-VQA methods results in missing high-frequency details for many videos. In this paper, we propose a novel Transformer-based NR-VQA framework that preserves the high-resolution quality information. With the multi-resolution input representation and a novel multi-resolution patch sampling mechanism, our method enables a comprehensive view of both the global video composition and local high-resolution details. The proposed approach can effectively aggregate quality information across different granularities in spatial and temporal dimensions, making the model robust to input resolution variations. Our method achieves state-of-the-art performance on large-scale UGC VQA datasets LSVQ and LSVQ-1080p, and on KoNViD-1k and LIVE-VQC without fine-tuning.
Reference35 articles.
1. Quantifying attention flow in transformers;Abnar,2020
2. Vivit: A video vision transformer;Arnab,2021
3. Longformer: The long-document transformer;Beltagy,2020
4. Is space-time attention all you need for video understanding?;Bertasius;Int. Conf. Mach. Learn. (ICML),2021
5. End-to-end object detection with transformers;Carion,2020
Cited by
3 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献