Affiliation:
1. Heilongjiang Province Key Laboratory of Laser Spectroscopy Technology and Application, Harbin University of Science and Technology, Harbin 150080, China
2. Department of Computer Science, Chubu University, Kasugai 487-8501, Japan
3. College of Electron and Information, University of Electronic Science and Technology of China, Zhongshan Institute, Zhongshan 528402, China
Abstract
To address the issue of reduced gaze estimation accuracy caused by individual differences in different environments, this study proposes a novel gaze estimation algorithm based on attention mechanisms. Firstly, by constructing a facial feature extractor (FFE), the method obtains facial feature information about the eyes and locates the feature areas of the left and right eyes. Then, the L2CSNet (l2 loss + cross-entropy loss + softmax layer network), which integrates the PSA (pyramid squeeze attention), is designed to increase the correlation weights related to gaze estimation in the feature areas, suppress other irrelevant weights, and extract more fine-grained feature information to obtain gaze direction features. Finally, by integrating L2CSNet with FFE and PSA, FPSA_L2CSNet was proposed, which is fully tested on four representative publicly available datasets and a real-world dataset comprising individuals of different backgrounds, lighting conditions, nationalities, skin tones, ages, genders, and partial occlusions. The experimental results indicate that the accuracy of the gaze estimation model proposed in this paper has been improved by 13.88%, 11.43%, and 7.34%, compared with L2CSNet, FSE_L2CSNet, and FCBA_L2CSNet, respectively. This model not only improves the robustness of gaze estimation but also provides more accurate estimation results than the original model.
Funder
Major Science and Technology Projects of Zhongshan City
Subject
Electrical and Electronic Engineering,Computer Networks and Communications,Hardware and Architecture,Signal Processing,Control and Systems Engineering
Reference48 articles.
1. Toward Mobile Eye-Based Human-Computer Interaction;Bulling;IEEE Pervasive Comput.,2010
2. Drewes, H. (2010). Eye gaze Tracking for Human Computer Interaction. [Ph.D. Thesis, LMU].
3. Patney, A., Kim, J., and Salvi, M. (2016, January 24–28). Perceptually-based foveated virtual reality. Proceedings of the ACM SIGGRAPH 2016 Emerging Technologies, Anaheim, CA, USA.
4. Novak, D., and Riener, R. (2013, January 24–26). Enhancing patient freedom in rehabilitation robotics using gaze-based intention detection. Proceedings of the IEEE International Conference on Rehabilitation Robotics, Seattle, WA, USA.
5. What do surgeons see: Capturing and synchronizing eye gaze for surgery applications;Atkins;Surg. Innov.,2013
Cited by
1 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献