Learning Adaptive Quantization Parameter for Consistent Quality Oriented Video Coding
-
Published:2023-12-06
Issue:24
Volume:12
Page:4905
-
ISSN:2079-9292
-
Container-title:Electronics
-
language:en
-
Short-container-title:Electronics
Author:
Vu Tien Huu1, Do Minh Ngoc2, Nguyen Sang Quang2, PhiCong Huy1ORCID, Sisouvong Thipphaphone3, HoangVan Xiem2
Affiliation:
1. Faculty of Multimedia, Posts and Telecommunications Institute of Technology, Hanoi 100000, Vietnam 2. Faculty of Electronics and Telecommunications, VNU—University of Engineering and Technology, Vietnam National University, Hanoi 100000, Vietnam 3. Faculty of Postgraduate Studies, Posts and Telecommunications Institute of Technology, Hanoi 100000, Vietnam
Abstract
In the industry 4.0 era, video applications such as surveillance visual systems, video conferencing, or video broadcasting have been playing a vital role. In these applications, for manipulating and tracking objects in decoded video, the quality of decoded video should be consistent because it largely affects the performance of the machine analysis. To cope with this problem, we propose a novel perceptual video coding (PVC) solution in which a full reference quality metric named video multimethod assessment fusion (VMAF) is employed together with a deep convolutional neural network (CNN) to obtain consistent quality while still achieving high compression performance. First of all, in order to achieve the consistent quality requirement, we propose a CNN model with an expected VMAF as input to adaptively adjust the quantization parameters (QP) for each coding block. Afterwards, to increase the compression performance, a Lagrange coefficient of rate-distortion optimization (RDO) mechanism is adaptively computed according to rate-QP and quality-QP models. The experimental results show that the proposed PVC solution has achieved two targets simultaneously: the quality of video sequence is kept consistently with an expected quality level and the bit rate saving of the proposed method is higher than traditional video coding standards and the relevant benchmark, notably with around 10% bitrate saving on average.
Funder
Research Collaboration Project between PTIT and Naver Corp.
Subject
Electrical and Electronic Engineering,Computer Networks and Communications,Hardware and Architecture,Signal Processing,Control and Systems Engineering
Reference35 articles.
1. Brunnström, K., Beker, S.A., de Moor, K., Dooms, A., Egger, S., Garcia, M.-N., Hossfeld, T., Jumisko-Pyykkö, S., Keimel, C., and Larabi, M.-C. (2023, August 10). Qualinet White Paper on Definitions of Quality of Experience. Available online: https://hal.science/hal-00977812/document. 2. Hoßfeld, T., Seufert, M., Sieber, C., and Zinner, T. (2014, January 18–20). Assessing effect sizes of influence factors towards a QoE model for HTTP adaptive streaming. Proceedings of the 2014 Sixth International Workshop on Quality of Multimedia Experience (QoMEX), Singapore. 3. A quality-of-content-based joint source and channel coding for human detections in a mobile surveillance cloud;Chen;IEEE Trans. Circuits Syst. Video Technol.,2017 4. Milani, S., Bernardini, R., and Rinaldo, R. (2013, January 26–31). A saliency-based rate control for people detection in video. Proceedings of the 2013 IEEE International Conference on Acoustics, Speech and Signal Processing, Vancouver, BC, Canada. 5. Low-pass filtering of rate-distortion functions for quality smoothing in real-time video communication;He;IEEE Trans. Circuits Syst. Video Technol.,2005
|
|