Author:
Du Tongchun,Wang Bo,Hu Liangchen
Abstract
Abstract
Deep reinforcement learning is a data-driven method, which is very promising for alleviating traffic congestion through intelligent control of traffic lights. In this paper, the traffic signal of an intersection is divided into four independent phases and then controlled by deep Q-network (DQN) models respectively. Models can receive observations from their own angle of view, i.e., north-south straight, north-south left turn, east-west straight, east-west left turn, instead of extracting features from the whole scene. We suppose that it is beneficial for learning better policy if agents could sense the environment more precisely. DQN models are jointly trained under the revised QMIX framework to promote coordination capability. For decentralized execution, traffic lights of the phase with the highest Q-value will turn green. The experiments are done under SUMO, the results demonstrate that our method obtains higher reward and lower delay compared to controlling the holistic cycle by using a single DQN model.
Subject
Computer Science Applications,History,Education
Reference37 articles.
1. Monotonic Value Function Factorisation for Deep Multi- Agent Reinforcement Learning;Rashid;Journal of Machine Learning Research,2018
2. A Survey and Critque of Multiagent Deep Reinforcement Learning;Hernandez-Leal;Autonomous Agents and Multi-Agent Systems,2019
3. AVD-Net: Attention value decomposition network for deep multi-agent reinforcement learning;Zhang,2021
Cited by
2 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献