Affiliation:
1. School of Integrated Technology, Yonsei University, Seodaemun-Gu, Seoul, South Korea
2. School of Information Systems, Singapore Management University, Singapore
Abstract
Gaze tracking is a key building block used in many mobile applications including entertainment, personal productivity, accessibility, medical diagnosis, and visual attention monitoring. In this paper, we present iMon, an appearance-based gaze tracking system that is both designed for use on mobile phones and has significantly greater accuracy compared to prior state-of-the-art solutions. iMon achieves this by comprehensively considering the gaze estimation pipeline and then overcoming three different sources of errors. First, instead of assuming that the user's gaze is fixed to a single 2D coordinate, we construct each gaze label using a probabilistic 2D heatmap gaze representation input to overcome errors caused by microsaccade eye motions that cause the exact gaze point to be uncertain. Second, we design an image enhancement model to refine visual details and remove motion blur effects of input eye images. Finally, we apply a calibration scheme to correct for differences between the perceived and actual gaze points caused by individual Kappa angle differences. With all these improvements, iMon achieves a person-independent per-frame tracking error of 1.49 cm (on smartphones) and 1.94 cm (on tablets) when tested with the GazeCapture dataset and 2.01 cm with the TabletGaze dataset. This outperforms the previous state-of-the-art solutions by ~22% to 28%. By averaging multiple per-frame estimations that belong to the same fixation point and applying personal calibration, the tracking error is further reduced to 1.11 cm (smartphones) and 1.59 cm (tablets). Finally, we built implementations that run on an iPhone 12 Pro and show that our mobile implementation of iMon can run at up to 60 frames per second - thus making gaze-based control of applications possible.
Funder
National Research Foundation of Korea
Publisher
Association for Computing Machinery (ACM)
Subject
Computer Networks and Communications,Hardware and Architecture,Human-Computer Interaction
Reference66 articles.
1. Jason Antic Jeremy Howard and Uri Manor. 2019. Decrappification DeOldification and Super Resolution. https://www.fast.ai/2019/05/03/decrappify/. Jason Antic Jeremy Howard and Uri Manor. 2019. Decrappification DeOldification and Super Resolution. https://www.fast.ai/2019/05/03/decrappify/.
2. Saeed Anwar , Salman Khan , and Nick Barnes . 2019. A deep journey into super-resolution: A survey. arXiv preprint arXiv:1904.07523 ( 2019 ). Saeed Anwar, Salman Khan, and Nick Barnes. 2019. A deep journey into super-resolution: A survey. arXiv preprint arXiv:1904.07523 (2019).
3. Constrained Local Neural Fields for Robust Facial Landmark Detection in the Wild
4. Adaptive Feature Fusion Network for Gaze Tracking in Mobile Tablets
5. G. Bradski . 2000. The OpenCV Library. Dr. Dobb's Journal of Software Tools ( 2000 ). G. Bradski. 2000. The OpenCV Library. Dr. Dobb's Journal of Software Tools (2000).
Cited by
11 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献