Interactive task learning via embodied corrective feedback
-
Published:2020-09-27
Issue:2
Volume:34
Page:
-
ISSN:1387-2532
-
Container-title:Autonomous Agents and Multi-Agent Systems
-
language:en
-
Short-container-title:Auton Agent Multi-Agent Syst
Author:
Appelgren MattiasORCID, Lascarides Alex
Abstract
AbstractThis paper addresses a task in Interactive Task Learning (Laird et al. IEEE Intell Syst 32:6–21, 2017). The agent must learn to build towers which are constrained by rules, and whenever the agent performs an action which violates a rule the teacher provides verbal corrective feedback: e.g. “No, red blocks should be on blue blocks”. The agent must learn to build rule compliant towers from these corrections and the context in which they were given. The agent is not only ignorant of the rules at the start of the learning process, but it also has a deficient domain model, which lacks the concepts in which the rules are expressed. Therefore an agent that takes advantage of the linguistic evidence must learn the denotations of neologisms and adapt its conceptualisation of the planning domain to incorporate those denotations. We show that by incorporating constraints on interpretation that are imposed by discourse coherence into the models for learning (Hobbs in On the coherence and structure of discourse, Stanford University, Stanford, 1985; Asher et al. in Logics of conversation, Cambridge University Press, Cambridge, 2003), an agent which utilizes linguistic evidence outperforms a strong baseline which does not.
Funder
Engineering and Physical Sciences Research Council University of Edinburgh
Publisher
Springer Science and Business Media LLC
Subject
Artificial Intelligence
Reference67 articles.
1. Afantenos, S. D., Kow, E., Asher, N. & Perret, J. (2015). Discourse parsing for multi-party chat dialogues. In L. Màrquez, C. Callison-Burch, J. Su., D. Pighin & Y. Marton (Eds.), Proceedings of the 2015 conference on empirical methods in natural language processing, EMNLP 2015, Lisbon, Portugal, September 17-21, 2015, The Association for Computational Linguistics (pp. 928–937) https://doi.org/10.18653/v1/d15-1109. 2. Al-Omari, M., Duckworth, P., Hawasly, M., Hogg, D. C. & Cohn, A. G. (2017). Natural language grounding and grammar induction for robotic manipulation commands. In M. Bansal, C. Matuszek, J. Andreas, Y. Artzi & Y. Bisk (Eds.), Proceedings of the first workshop on language grounding for robotics, RoboNLP@ACL 2017, Vancouver, Canada, August 3, 2017, Association for Computational Linguistics (pp. 35–43) https://doi.org/10.18653/v1/w17-2805. 3. Anderson, P., Wu, Q., Teney, D., Bruce, J., Johnson, M., Sünderhauf, N., Reid, I. D., Gould, S. & van den Hengel, A. (2017). Vision-and-language navigation: Interpreting visually-grounded navigation instructions in real environments. In 2018 IEEE/CVF conference on computer vision and pattern recognition (pp. 3674–3683). 4. Appelgren, M. & Lascarides, A. (2019a). Coherence, symbol grounding and interactive task learning. In Proceedings of the 23rd workshop on the semantics and pragmatics of dialogue: full papers, SEMDIAL, London, United Kingdom. http://semdial.org/anthology/Z19-Appelgren_semdial_0004.pdf. 5. Appelgren, M. & Lascarides, A. (2019b). Learning plans by acquiring grounded linguistic meanings from corrections. In Proceedings of the 18th inter- national conference on autonomous agents and multiagent systems (AAMAS 2019) Montreal, Canada, May 13–17, 2019, IFAAMAS (p. 9).
Cited by
3 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
|
|