Affiliation:
1. New York University and The Chinese University of Hong Kong, Hong Kong SAR
2. New York University
3. The Chinese University of Hong Kong, Hong Kong SAR and Shanghai Artificial Intelligence Laboratory, China
Abstract
While tremendous advances in visual and auditory realism have been made for virtual and augmented reality (VR/AR), introducing a plausible sense of physicality into the virtual world remains challenging. Closing the gap between real-world physicality and immersive virtual experience requires a closed interaction loop: applying user-exerted physical forces to the virtual environment and generating haptic sensations back to the users. However, existing VR/AR solutions either completely ignore the force inputs from the users or rely on obtrusive sensing devices that compromise user experience.
By identifying users' muscle activation patterns while engaging in VR/AR, we design a learning-based neural interface for natural and intuitive force inputs. Specifically, we show that lightweight electromyography sensors, resting non-invasively on users' forearm skin, inform and establish a robust understanding of their complex hand activities. Fuelled by a neural-network-based model, our interface can decode finger-wise forces in real-time with 3.3% mean error, and generalize to new users with little calibration. Through an interactive psychophysical study, we show that human perception of virtual objects' physical properties, such as stiffness, can be significantly enhanced by our interface. We further demonstrate that our interface enables ubiquitous control via finger tapping. Ultimately, we envision our findings to push forward research towards more realistic physicality in future VR/AR.
Publisher
Association for Computing Machinery (ACM)
Subject
Computer Graphics and Computer-Aided Design
Reference112 articles.
1. A Ubiquitous WiFi-Based Fine-Grained Gesture Recognition System
2. WiGest: A ubiquitous WiFi-based gesture recognition system
3. EMG signal classification for human computer interaction: a review;Ahsan Md Rezwanul;European Journal of Scientific Research,2009
4. Journal of Physics: Conference Series;Andrean Deni,2030
5. Gopala K Anumanchipalli , Josh Chartier , and Edward F Chang . 2019. Speech synthesis from neural decoding of spoken sentences. Nature 568, 7753 ( 2019 ), 493--498. Gopala K Anumanchipalli, Josh Chartier, and Edward F Chang. 2019. Speech synthesis from neural decoding of spoken sentences. Nature 568, 7753 (2019), 493--498.
Cited by
8 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献