Affiliation:
1. Beijing Engineering Research Center of Mixed Reality and Advanced Display, School of Optics and Photonics, Beijing Institute of Technology, Beijing 100081, China
Abstract
Multimodal interaction systems can provide users with natural and compelling interactive experiences. Despite the availability of various sensing devices, only some commercial multimodal applications are available. One reason may be the need for a more efficient framework for fusing heterogeneous data and addressing resource pressure. This paper presents a parallel multimodal integration framework that ensures that the errors and external damages of integrated devices remain uncorrelated. The proposed relative weighted fusion method and modality delay strategy process the heterogeneous data at the decision level. The parallel modality operation flow allows each device to operate across multiple terminals, reducing resource demands on a single computer. The universal fusion methods and independent devices further remove constraints on the integrated modality number, providing the framework with extensibility. Based on the framework, we develop a multimodal virtual shopping system, integrating five input modalities and three output modalities. The objective experiments show that the system can accurately fuse heterogeneous data and understand interaction intent. User studies indicate the immersive and entertaining of multimodal shopping. Our framework proposes a development paradigm for multimodal systems, fostering multimodal applications across various domains.
Funder
National Key R&D Program of China
2022 major science and technology project “Yuelu·Multimodal Graph-Text-Sound-Semantic Gesture Big Model Research and Demonstration Application“ in Changsha
Strategic research and consulting project of Chinese Academy of Engineering
Subject
Fluid Flow and Transfer Processes,Computer Science Applications,Process Chemistry and Technology,General Engineering,Instrumentation,General Materials Science
Reference56 articles.
1. Multimodal human-computer interaction: A survey;Jaimes;Comput. Vis. Image Underst.,2007
2. Dumas, B., Lalanne, D., and Oviatt, S. (2009). Human Machine Interaction: Research Results of the Mmi Program, Springer.
3. Multimodal interaction: A review;Turk;Pattern Recognit. Lett.,2014
4. Flippo, F., Krebs, A., and Marsic, I. (2003, January 5–7). A Framework for Rapid Development of Multimodal Interfaces. Proceedings of the 5th International Conference on Multimodal Interfaces, ICMI ’03, Vancouver, BC, Canada.
5. A data-level fusion approach for degradation modeling and prognostic analysis under multiple failure modes;Abdallah;J. Qual. Technol.,2018