Abstract
A haptic sensor coupled to a gamepad or headset is frequently used to enhance the sense of immersion for game players. However, providing haptic feedback for appropriate sound effects involves specialized audio engineering techniques to identify target sounds that vary according to the game. We propose a deep learning-based method for sound event detection (SED) to determine the optimal timing of haptic feedback in extremely noisy environments. To accomplish this, we introduce the BattleSound dataset, which contains a large volume of game sound recordings of game effects and other distracting sounds, including voice chats from a PlayerUnknown’s Battlegrounds (PUBG) game. Given the highly noisy and distracting nature of war-game environments, we set the annotation interval to 0.5 s, which is significantly shorter than the existing benchmarks for SED, to increase the likelihood that the annotated label contains sound from a single source. As a baseline, we adopt mobile-sized deep learning models to perform two tasks: weapon sound event detection (WSED) and voice chat activity detection (VCAD). The accuracy of the models trained on BattleSound was greater than 90% for both tasks; thus, BattleSound enables real-time game sound recognition in noisy environments via deep learning. In addition, we demonstrated that performance degraded significantly when the annotation interval was greater than 0.5 s, indicating that the BattleSound with short annotation intervals is advantageous for SED applications that demand real-time inferences.
Funder
Korea Institute of Energy Technology Evaluation and Planning
Subject
Electrical and Electronic Engineering,Biochemistry,Instrumentation,Atomic and Molecular Physics, and Optics,Analytical Chemistry
Reference54 articles.
1. Use of auditory event-related potentials to measure immersion during a computer game;Burns;Int. J. Hum. Comput. Stud.,2015
2. Stach, T., and Graham, T.C.N. (2011, January 5–9). Exploring Haptic Feedback in Exergames. Proceedings of the 13th IFIP TC 13 International Conference on Human-Computer Interaction—Volume Part II, Lisbon, Portugal. INTERACT’11.
3. A comparison of the effects of haptic and visual feedback on presence in virtual reality;Gibbs;Int. J. Hum.-Comput. Stud.,2022
4. Game audio—An investigation into the effect of audio on player immersion;Gallacher;Comput. Games J.,2013
5. A statistical model-based voice activity detection;Sohn;IEEE Signal Process. Lett.,1999