Affiliation:
1. Univ. Rennes, INSA Rennes, CNRS, IETR - UMR 6164, France
2. National Higher School of Telecommunications and ICT, Algeria
Abstract
Recently, with the growing popularity of mobile devices as well as video sharing platforms (e.g., YouTube, Facebook, TikTok, and Twitch), User-Generated Content (UGC) videos have become increasingly common and now account for a large portion of multimedia traffic on the internet. Unlike professionally generated videos produced by filmmakers and videographers, typically, UGC videos contain multiple authentic distortions, generally introduced during capture and processing by naive users. Quality prediction of UGC videos is of paramount importance to optimize and monitor their processing in hosting platforms, such as their coding, transcoding, and streaming. However, blind quality prediction of UGC is quite challenging, because the degradations of UGC videos are unknown and very diverse, in addition to the unavailability of pristine reference. Therefore, in this article, we propose an accurate and efficient Blind Video Quality Assessment (BVQA) model for UGC videos, which we name 2BiVQA for double Bi-LSTM Video Quality Assessment. 2BiVQA metric consists of three main blocks, including a pre-trained Convolutional Neural Network to extract discriminative features from image patches, which are then fed into two Recurrent Neural Networks for spatial and temporal pooling. Specifically, we use two Bi-directional Long Short-term Memory networks, the first is used to capture short-range dependencies between image patches, while the second allows capturing long-range dependencies between frames to account for the temporal memory effect. Experimental results on recent large-scale UGC VQA datasets show that 2BiVQA achieves high performance at lower computational cost than most state-of-the-art VQA models. The source code of our 2BiVQA metric is made publicly available at
https://github.com/atelili/2BiVQA
.
Publisher
Association for Computing Machinery (ACM)
Subject
Computer Networks and Communications,Hardware and Architecture
Reference105 articles.
1. Sewoong Ahn and Sanghoon Lee. 2018. Deep blind video quality assessment based on temporal human perception. In Proceedings of the 25th IEEE International Conference on Image Processing (ICIP’18). IEEE, 619–623.
2. Image quality assessment by comparing CNN features between images;Amirshahi S. A.;J. Imag. Sci. Technol.,2016
3. Alexey Bochkovskiy Chien-Yao Wang and Hong-Yuan Mark Liao. 2020. YOLOv4: Optimal speed and accuracy of object detection. Retrieved from https://arxiv.org/abs/2004.10934
4. ITU Recommendation BT. 2012. Methodology for the Subjective Assessment of the Quality of Television Pictures. Int. Telecommun. Union 6 (2012).
Cited by
2 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献