Distributed agent-based deep reinforcement learning for large scale traffic signal control

作者:

Highlights:

摘要

Traffic signal control (TSC) is an established yet challenging engineering solution that alleviates traffic congestion by coordinating vehicles’ movements at road intersections. Theoretically, reinforcement learning (RL) is a promising method for adaptive TSC in complex urban traffic networks. However, current TSC systems still rely heavily on simplified rule-based methods in practice. In this paper, we propose: (1) two game theory-aided RL algorithms leveraging Nash Equilibrium and RL, namely Nash Advantage Actor–Critic (Nash-A2C) and Nash Asynchronous Advantage Actor–Critic (Nash-A3C); (2) a distributed computing Internet of Things (IoT) architecture for traffic simulation, which is more suitable for distributed TSC methods like the Nash-A3C deployment in its fog layer. We apply both methods in our computing architecture and obtain better performance than benchmark TSC methods by 22.1% and 9.7% reduction of congestion time and network delay, respectively.

论文关键词:Traffic signal control,Reinforcement learning,Nash Equilibrium,Nash-A3C,Distributed computing architecture

论文评审过程:Received 3 November 2021, Revised 11 January 2022, Accepted 24 January 2022, Available online 3 February 2022, Version of Record 12 February 2022.

论文官网地址:https://doi.org/10.1016/j.knosys.2022.108304