A Light-Field Video Dataset of Scenes with Moving Objects Captured with a Plenoptic Video Camera
-
Published:2024-06-06
Issue:11
Volume:13
Page:2223
-
ISSN:2079-9292
-
Container-title:Electronics
-
language:en
-
Short-container-title:Electronics
Author:
Javidi Kamran1ORCID, Martini Maria G.1ORCID
Affiliation:
1. Wireless and Multimedia Networking Research Group, Department of Networks and Digital Media, School of Computer Science and Mathematics, Faculty of Engineering, Computing and the Environment, Kingston University, Penrhyn Road Campus, Kingston upon Thames KT1 2EE, UK
Abstract
Light-field video provides a detailed representation of scenes captured from different perspectives. This results in a visualisation modality that enhances the immersion and engagement of the viewers with the depicted environment. In order to perform research on compression, transmission and signal processing of light field data, datasets with light-field contents of different categories and acquired with different modalities are required. In particular, the development of machine learning models for quality assessment and for light-field processing, including the generation of new views, require large amounts of data. Most existing datasets consist of static scenes and, in many cases, synthetic contents. This paper presents a novel light-field plenoptic video dataset, KULFR8, involving six real-world scenes with moving objects and 336 distorted light-field videos derived from the original contents; in total, the original scenes in the dataset contain 1800 distinctive frames, with angular resolution of 5×5 with and total spatial resolution of 9600×5400 pixels (considering all the views); overall, the dataset consists of 45,000 different views with spatial resolution of 1920×1080 pixels. We analyse the content characteristics based on the dimensions of the captured objects and via the acquired videos using the central views extracted from each quilted frame. Additionally, we encode and decode the contents using various video encoders across different bitrate ranges. For quality assessments, we consider all the views, utilising frames measuring 9600×5400 pixels, and employ two objective quality metrics: PSNR and SSIM.
Reference59 articles.
1. Light field image processing: An overview;Wu;IEEE J. Sel. Top. Signal Process.,2017 2. Levoy, M., and Hanrahan, P. (1996, January 4–9). Light field rendering. Proceedings of the 23rd Annual Conference on Computer Graphics and Interactive Techniques, New Orleans, LA, USA. 3. Wang, Y., Wang, L., Liang, Z., Yang, J., Timofte, R., Guo, Y., Jin, K., Wei, Z., Yang, A., and Guo, S. (2023, January 18–22). NTIRE 2023 challenge on light field image super-resolution: Dataset, methods and results. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada. 4. A multispectral light field dataset and framework for light field deep learning;Schambach;IEEE Access,2020 5. Jin, J., Hou, J., Chen, J., and Kwong, S. (2020, January 13–19). Light field spatial super-resolution via deep combinatorial geometry embedding and structural consistency regularization. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA.
|
|