Affiliation:
1. Technical University of Crete, Chania - Crete, Greece
2. Durham University, Durham, United Kingdom
Abstract
Input methods for interaction in smartphone-based virtual and mixed reality (VR/MR) are currently based on uncomfortable head tracking controlling a pointer on the screen. User fixations are a fast and natural input method for VR/MR interaction. Previously, eye tracking in mobile VR suffered from low accuracy, long processing time, and the need for hardware add-ons such as anti-reflective lens coating and infrared emitters. We present an innovative mobile VR eye tracking methodology utilizing only the eye images from the front-facing (selfie) camera through the headset’s lens, without any modifications. Our system first enhances the low-contrast, poorly lit eye images by applying a pipeline of customised low-level image enhancements suppressing obtrusive lens reflections. We then propose an iris region-of-interest detection algorithm that is run only once. This increases the iris tracking speed by reducing the iris search space in mobile devices. We iteratively fit a customised geometric model to the iris to refine its coordinates. We display a thin bezel of light at the top edge of the screen for constant illumination. A confidence metric calculates the probability of successful iris detection. Calibration and linear gaze mapping between the estimated iris centroid and physical pixels on the screen results in low latency, real-time iris tracking. A formal study confirmed that our system’s accuracy is similar to eye trackers in commercial VR headsets in the central part of the headset’s field-of-view. In a VR game, gaze-driven user completion time was as fast as with head-tracked interaction, without the need for consecutive head motions. In a VR panorama viewer, users could successfully switch between panoramas using gaze.
Publisher
Association for Computing Machinery (ACM)
Subject
Experimental and Cognitive Psychology,General Computer Science,Theoretical Computer Science
Reference44 articles.
1. [n.d.]. OpenCV Library. Retrieved May 1 2020 from https://opencv.org/. [n.d.]. OpenCV Library. Retrieved May 1 2020 from https://opencv.org/.
2. [n.d.]. XCode. Retrieved September 9 2019 from https://developer.apple.com/xcode/. [n.d.]. XCode. Retrieved September 9 2019 from https://developer.apple.com/xcode/.
3. EyeSpyVR: Interactive eye sensing using off-the-shelf, smartphone-based VR headsets;Ahuja Karan;Proc. ACM Interact. Mobile Wearable Ubiq. Technol.,2018
4. Quantitative measurement of saccade amplitude, duration, and velocity
Cited by
17 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献