Abstract
AbstractGaze control represents an important issue in the interaction between a robot and humans. Specifically, deciding who to pay attention to in a multi-party conversation is one way to improve the naturalness of a robot in human-robot interaction. This control can be carried out by means of two different models that receive the stimuli produced by the participants in an interaction, either an on-center off-surround competitive network or a recurrent neural network. A system based on a competitive neural network is able to decide who to look at with a smooth transition in the focus of attention when significant changes in stimuli occur. An important aspect in this process is the configuration of the different parameters of such neural network. The weights of the different stimuli have to be computed to achieve human-like behavior. This article explains how these weights can be obtained by solving an optimization problem. In addition, a new model using a recurrent neural network with LSTM layers is presented. This model uses the same set of stimuli but does not require its weighting. This new model is easier to train, avoiding manual configurations, and offers promising results in robot gaze control. The experiments carried out and some results are also presented.
Funder
Ministerio de Ciencia, Innovación y Universidades
Programa de Apoyo a Proyectos de Investigación de la Junta de Castilla y León
Publisher
Springer Science and Business Media LLC
Subject
Computer Networks and Communications,Hardware and Architecture,Media Technology,Software
Reference42 articles.
1. Abd El-Moneim S, Nassar M, Dessouky MI, Ismail NA, El-Fishawy AS, Abd El-Samie FE (2020) Text-independent speaker recognition using lstm-rnn and speech enhancement. Mult Tools Appl 79(33):24,013–24,028
2. Admoni H, Scassellati B (2017) Social eye gaze in human-robot interaction: a review. J Human Robot Interact 6(1):25–63
3. Alonso-Martín F, Gorostiza JF, Malfaz M, Salichs MA (2012) User localization during human-robot interaction. Sensors 12(7):9913–9935
4. Andrist S, Mutlu B, Tapus A (2015) Look like me: matching robot personality via gaze to increase motivation. In: Proceedings of the 33rd annual ACM conference on human factors in computing systems, pp 3603–3612. ACM
5. Bendris M, Charlet D, Chollet G (2010) Lip activity detection for talking faces classification in tv-content. In: International conference on machine vision, pp 187–190
Cited by
7 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献