Dopamine encoding of novelty facilitates efficient uncertainty-driven exploration

Author:

Wang YuhaoORCID,Lak Armin,Manohar Sanjay G.,Bogacz RafalORCID

Abstract

When facing an unfamiliar environment, animals need to explore to gain new knowledge about which actions provide reward, but also put the newly acquired knowledge to use as quickly as possible. Optimal reinforcement learning strategies should therefore assess the uncertainties of these action–reward associations and utilise them to inform decision making. We propose a novel model whereby direct and indirect striatal pathways act together to estimate both the mean and variance of reward distributions, and mesolimbic dopaminergic neurons provide transient novelty signals, facilitating effective uncertainty-driven exploration. We utilised electrophysiological recording data to verify our model of the basal ganglia, and we fitted exploration strategies derived from the neural model to data from behavioural experiments. We also compared the performance of directed exploration strategies inspired by our basal ganglia model with other exploration algorithms including classic variants of upper confidence bound (UCB) strategy in simulation. The exploration strategies inspired by the basal ganglia model can achieve overall superior performance in simulation, and we found qualitatively similar results in fitting model to behavioural data compared with the fitting of more idealised normative models with less implementation level detail. Overall, our results suggest that transient dopamine levels in the basal ganglia that encode novelty could contribute to an uncertainty representation which efficiently drives exploration in reinforcement learning.

Funder

Biotechnology and Biological Sciences Research Council

Medical Research Council

Wellcome Trust

Royal Society

National Institute for Healthcare Research

James S. McDonnell Foundation

Publisher

Public Library of Science (PLoS)

Reference49 articles.

1. Learning the payoffs and costs of actions;M Möller;PLOS Computational Biology,2019

2. Restless Bandits: Activity Allocation in a Changing World;P Whittle;Journal of Applied Probability,1988

3. Humans use directed and random exploration to solve the explore–exploit dilemma;RC Wilson;Journal of Experimental Psychology: General,2014

4. Bandit Processes and Dynamic Allocation Indices;JC Gittins;Journal of the Royal Statistical Society: Series B (Methodological),1979

5. The Multi-Armed Bandit Problem: Decomposition and Computation;MN Katehakis;Mathematics of Operations Research,1987

Cited by 1 articles. 订阅此论文施引文献 订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3