Traffic Signal Priority Control Based on Shared Experience Multi-Agent Deep Reinforcement Learning

Deep Reinforcement Learning (DRL) has demonstrated its great potential for Adaptive Traffic Signal Control (ATSC) tasks at single-intersection. In the transportation network multi-agent environment, cooperative learning among multi-agents has become a hot research topic. Based on the distributed control model, this paper presents a hybrid reward function model for the dynamic density method of intersections, which emphasizes the priority of emergency vehicles (EMV) while maximizing the traffic efficiency of social vehicles, and solves the problem of sparse reward due to the ambiguous guidance relationship between the multi-agent Deep Reinforcement Learning (MDRL) state and reward function of the urban road network scenario. On the other hand, based on multi-agent A2C (MA2C) algorithm, this paper presents Shared Experience MA2C (SEMA2C) between agents. In the transportation network, each intersection represented by an agent has similar task objectives. SMEA2C algorithm takes the current agent as the main body of self-learning, and utilizes the principle of importance sampling to learn from the experience data of the agents located at adjacent intersections. The experimental results show that the proposed SEMA2C performs well in multi-agent traffic signal control tasks, and has greater advantages than the similar algorithms.

  • Record URL:
  • Availability:
  • Supplemental Notes:
    • Abstract reprinted with permission of the Institution of Engineering and Technology.
  • Authors:
    • Wang, Zhiwen
    • Yang, Kangkang
    • Li, Long
    • Lu, Yanrong
    • Tao, Yufei
  • Publication Date: 2023-7


  • English

Media Info

Subject/Index Terms

Filing Info

  • Accession Number: 01888201
  • Record Type: Publication
  • Files: TRIS
  • Created Date: Jul 20 2023 9:13AM