A Quality of Experience and Visual Attention Evaluation for 360° Videos with Non-spatial and Spatial Audio

Author:

Hirway Amit1ORCID,Qiao Yuansong2ORCID,Murray Niall1ORCID

Affiliation:

1. Department of Computer and Software Engineering, Technological University of the Shannon - Midlands Midwest, Athlone, Ireland

2. Software Research Institute, Technological University of the Shannon - Midlands Midwest, Athlone, Ireland

Abstract

This article presents the results of an empirical study that aimed to investigate the influence of various types of audio (spatial and non-spatial) on the user quality of experience (QoE) of and visual attention in 360° videos. The study compared the head pose, eye gaze, pupil dilations, heart rate, and subjective responses of 73 users who watched ten 360° videos with different sound configurations. The configurations evaluated were no sound; non-spatial (stereo) audio; and two spatial sound conditions (first- and third-order ambisonics). The videos covered various categories and presented both indoor and outdoor scenarios. The subjective responses were analyzed using an ANOVA (Analysis of Variance) to assess mean differences between sound conditions. Data visualization was also employed to enhance the interpretability of the results. The findings reveal diverse viewing patterns, physiological responses, and subjective experiences among users watching 360° videos with different sound conditions. Spatial audio, in particular third-order ambisonics, garnered heightened attention. This is evident in increased pupil dilation and heart rate. Furthermore, the presence of spatial audio led to more diverse head poses when sound sources were distributed across the scene. These findings have important implications for the development of effective techniques for optimizing processing, encoding, distributing, and rendering content in virtual reality (VR) and 360° videos with spatialized audio. These insights are also relevant in the creative realms of content design and enhancement. They provide valuable guidance on how spatial audio influences user attention, physiological responses, and overall subjective experiences. Understanding these dynamics can assist content creators and designers in crafting immersive experiences that leverage spatialized audio to captivate users, enhance engagement, and optimize the overall quality of VR and 360° video content. The dataset, scripts used for data collection, ffmpeg commands used for processing the videos, and the subjective questionnaire and its statistical analysis are publicly available.

Funder

Science Foundation Ireland

European Regional Development Fund

Horizon Europe Framework Program

Publisher

Association for Computing Machinery (ACM)

Reference54 articles.

1. R. Shafi W. Shuai and M. U. Younus. 2020. 360-degree video streaming: A survey of the state of the art. Symmetry 12 1491.

2. J. E. Hoffman. 2016. Visual attention and eye movements. Attention 119--153.

3. Laurent Itti. 2000. Models of bottom-up and top-down visual attention.

4. Visual Attention and Applications in Multimedia Technologies

5. S. Poeschl-Guenther, K. Wall, and N. Döring. 2013. Integration of spatial sound in immersive virtual environments: An experimental study on effects of spatial sound on presence. In Proceedings of the IEEE Virtual Reality, 129–130.

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3