Author:
Chandramouli Suyog,Shi Danqing,Putkonen Aini,De Peuter Sebastiaan,Zhang Shanshan,Jokinen Jussi,Howes Andrew,Oulasvirta Antti
Abstract
AbstractComputational rationality explains human behavior as arising due to the maximization of expected utility under the constraints imposed by the environment and limited cognitive resources. This simple assumption, when instantiated via partially observable Markov decision processes (POMDPs), gives rise to a powerful approach for modeling human adaptive behavior, within which a variety of internal models of cognition can be embedded. In particular, such an instantiation enables the use of methods from reinforcement learning (RL) to approximate the optimal policy solution to the sequential decision-making problems posed to the cognitive system in any given setting; this stands in contrast to requiring ad hoc hand-crafted rules for capturing adaptive behavior in more traditional cognitive architectures. However, despite their successes and promise for modeling human adaptive behavior across everyday tasks, computationally rational models that use RL are not easy to build. Being a hybrid of theoretical cognitive models and machine learning (ML) necessitates that model building take into account appropriate practices from both cognitive science and ML. The design of psychological assumptions and machine learning decisions concerning reward specification, policy optimization, parameter inference, and model selection are all tangled processes rife with pitfalls that can hinder the development of valid and effective models. Drawing from a decade of work on this approach, a workflow is outlined for tackling this challenge and is accompanied by a detailed discussion of the pros and cons at key decision points.
Funder
Finnish Center for Artificial Intelligence
Academy of Finland
Publisher
Springer Science and Business Media LLC
Reference115 articles.
1. Abel, D., Hershkowitz, D., & Littman, M. (2016). Near optimal behavior via approximate state abstraction. In M. F. Balcan, & K. Q. Weinberger (Eds.), Proceedings of The 33rd International conference on machine learning (Proceedings of Machine Learning Research, Vol. 48) (pp. 2915–2923). PMLR, New York, New York, USA.
2. Anderson, J. R. (1991). Is human cognition adaptive? Behavioral and brain sciences, 14(3), 471–485.
3. Anderson, J. R., Bothell, D., Lebiere, C., & Matessa, M. (1998). An integrated theory of list memory. Journal of Memory and Language, 38(4), 341–380.
4. Anderson, J. R., Matessa, M., & Lebiere, C. (1997). ACT-R: A theory of higher level cognition and its relation to visual attention. Human-Computer Interaction, 12(4), 439–462.
5. Arif, A. S., & Stuerzlinger, W. (2009). Analysis of text entry performance metrics. In 2009 IEEE Toronto international conference science and technology for humanity (TIC-STH) (pp. 100–105). IEEE, New York, NY.