Publication Type
Journal Article
Version
acceptedVersion
Publication Date
8-2021
Abstract
The adaptive traffic signal control (ATSC) problem can be modeled as a multiagent cooperative game among urban intersections, where intersections cooperate to counter the city's traffic conditions. Recently, reinforcement learning (RL) has achieved marked successes in managing sequential decision making problems, which motivates us to apply RL in the ATSC problem. One of the largest challenges of this problem is that the observation of intersection is typically partially observable, which limits the learning performance of RL algorithms. Considering the large scale of intersections in an urban traffic environment, we use independent RL to solve ATSC problem in this study. We model ATSC problem as a partially observable weak cooperative traffic model (PO-WCTM). Different from a traditional IRL task that averages the returns of all agents in fully cooperative games, the learning goal of each intersection in PO-WCTM is to reduce the cooperative difficulty of learning, which is also consistent with the traffic environment hypothesis. To achieve the optimal cooperative strategy of PO-WCTM, we propose an IRL algorithm called Cooperative Important Lenient Double DQN (CIL-DDQN), which extends Double DQN (DDQN) algorithm using two mechanisms: the forgetful experience mechanism and the lenient weight training mechanism. The former mechanism decreases the importance of experiences stored in the experience reply buffers, while the latter mechanism increases the weight experiences with high estimation and 'leniently' trains the DDQN neural network. Experiments in two real traffic scenarios and one simulated traffic scenarios show that, CIL-DDQN outperforms other methods in almost all performance indicators of ATSC.
Keywords
Multiagent learning, Independent reinforcement learning, Cooperative Markov game, Traffic signal control
Discipline
OS and Networks | Software Engineering
Research Areas
Software and Cyber-Physical Systems
Publication
IEEE Transactions on Vehicular Technology
Volume
70
Issue
8
First Page
7426
Last Page
7436
ISSN
0018-9545
Identifier
10.1109/TVT.2021.3090796
Publisher
Institute of Electrical and Electronics Engineers
Citation
ZHANG, Chengwei; JIN, Shan; XUE, Wanli; XIE, Xiaofei; CHEN, Shengyong; and CHEN, Rong.
Independent reinforcement learning for weakly cooperative multiagent traffic control problem. (2021). IEEE Transactions on Vehicular Technology. 70, (8), 7426-7436.
Available at: https://ink.library.smu.edu.sg/sis_research/7052
Creative Commons License
This work is licensed under a Creative Commons Attribution-NonCommercial-No Derivative Works 4.0 International License.