Appearance-based gaze estimation with feature fusion of multi-level information elements

Author:

Ren Zhonghe1ORCID,Fang Fengzhou12ORCID,Hou Gaofeng1ORCID,Li Zihao1,Niu Rui1

Affiliation:

1. State Key Laboratory of Precision Measuring Technology and Instruments, Laboratory of Micro/Nano Manufacturing Technology (MNMT), Tianjin University , Tianjin 300072 , China

2. Centre of Micro/Nano Manufacturing Technology (MNMT-Dublin), University College Dublin , Dublin 4 , Ireland

Abstract

Abstract Gaze estimation is a fundamental task in many applications of cognitive sciences, human–computer interaction, and robotics. The purely data-driven appearance-based gaze estimation methods may suffer from a lack of interpretability, which prevents their applicability to pervasive scenarios. In this study, a feature fusion method with multi-level information elements is proposed to improve the comprehensive performance of the appearance-based gaze estimation model. The multi-level feature extraction and expression are carried out from the originally captured images, and a multi-level information element matrix is established. A gaze conduction principle is formulated for reasonably fusing information elements from the established matrix. According to the gaze conduction principle along with the matrix, a multi-level information element fusion (MIEF) model for gaze estimation is proposed. Then, several input modes and network structures of the MIEF model are designed, and a series of grouping experiments are carried out on a small-scale sub-dataset. Furthermore, the optimized input modes and network structures of the MIEF model are selected for training and testing on the whole dataset to verify and compare model performance. Experimental results show that optimizing the feature combination in the input control module and fine-tuning the computational architecture in the feature extraction module can improve the performance of the gaze estimation model, which would enable the reduction of the model by incorporating the critical features and thus improve the performance and accessibility of the method. Compared with the reference baseline, the optimized model based on the proposed feature fusion method of multi-level information elements can achieve efficient training and improve the test accuracy in the verification experiment. The average error is 1.63 cm on phones on the GazeCapture dataset, which achieves comparable accuracy with state-of-the-art methods.

Funder

National Natural Science Foundation of China

Tianjin University

Publisher

Oxford University Press (OUP)

Subject

Computational Mathematics,Computer Graphics and Computer-Aided Design,Human-Computer Interaction,Engineering (miscellaneous),Modeling and Simulation,Computational Mechanics

Reference82 articles.

1. MIMETIC: Mobile encrypted traffic classification using multimodal deep learning;Aceto;Computer Networks,2019

2. Social eye gaze in human–robot interaction: A review;Admoni;Journal of Human–Robot Interaction,2017

3. Explainable artificial intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI;Arrieta;Information Fusion,2020

4. Adaptive feature fusion network for gaze tracking in mobile tablets;Bao,2021

5. Data-driven models in human neuroscience and neuroengineering;Brunton;Current Opinion in Neurobiology,2019

Cited by 5 articles. 订阅此论文施引文献 订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3