iCatcher+: Robust and Automated Annotation of Infants’ and Young Children’s Gaze Behavior From Videos Collected in Laboratory, Field, and Online Studies

Author:

Erel Yotam1,Shannon Katherine Adams2ORCID,Chu Junyi3,Scott Kim3ORCID,Kline Struhl Melissa3,Cao Peng4,Tan Xincheng5,Hart Peter3,Raz Gal3,Piccolo Sabrina3,Mei Catherine3,Potter Christine6,Jaffe-Dax Sagi7ORCID,Lew-Williams Casey8,Tenenbaum Joshua39,Fairchild Katherine9,Bermano Amit1,Liu Shari310ORCID

Affiliation:

1. The Blavatnik School of Computer Science, Tel Aviv University, Tel Aviv-Yafo, Israel

2. Department of Psychology, Stanford University, Palo Alto, California

3. Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, Cambridge, Massachusetts

4. Computer Science & Artificial Intelligence Laboratory, Massachusetts Institute of Technology, Cambridge, Massachusetts

5. School of Engineering and Applied Sciences, Harvard University, Cambridge, Massachusetts

6. Department of Psychology, The University of Texas at El Paso, El Paso, Texas

7. The School of Psychological Sciences, Tel Aviv University, Tel Aviv-Yafo, Israel

8. Department of Psychology, Princeton University, Princeton, New Jersey

9. The MIT Quest for Intelligence, Massachusetts Institute of Technology, Cambridge, Massachusetts

10. Department of Psychological and Brain Sciences, Johns Hopkins University, Baltimore, Maryland

Abstract

Technological advances in psychological research have enabled large-scale studies of human behavior and streamlined pipelines for automatic processing of data. However, studies of infants and children have not fully reaped these benefits because the behaviors of interest, such as gaze duration and direction, still have to be extracted from video through a laborious process of manual annotation, even when these data are collected online. Recent advances in computer vision raise the possibility of automated annotation of these video data. In this article, we built on a system for automatic gaze annotation in young children, iCatcher, by engineering improvements and then training and testing the system (referred to hereafter as iCatcher+) on three data sets with substantial video and participant variability (214 videos collected in U.S. lab and field sites, 143 videos collected in Senegal field sites, and 265 videos collected via webcams in homes; participant age range = 4 months–3.5 years). When trained on each of these data sets, iCatcher+ performed with near human-level accuracy on held-out videos on distinguishing “LEFT” versus “RIGHT” and “ON” versus “OFF” looking behavior across all data sets. This high performance was achieved at the level of individual frames, experimental trials, and study videos; held across participant demographics (e.g., age, race/ethnicity), participant behavior (e.g., movement, head position), and video characteristics (e.g., luminance); and generalized to a fourth, entirely held-out online data set. We close by discussing next steps required to fully automate the life cycle of online infant and child behavioral studies, representing a key step toward enabling robust and high-throughput developmental research.

Publisher

SAGE Publications

Subject

General Psychology

Cited by 5 articles. 订阅此论文施引文献 订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3