Proceedings of the 28th ACM International Conference on Information and Knowledge Management 2019
DOI: 10.1145/3357384.3357900
|View full text |Cite
|
Sign up to set email alerts
|

Learning Phase Competition for Traffic Signal Control

Abstract: Increasingly available city data and advanced learning techniques have empowered people to improve the efficiency of our city functions. Among them, improving the urban transportation efficiency is one of the most prominent topics. Recent studies have proposed to use reinforcement learning (RL) for traffic signal control. Different from traditional transportation approaches which rely heavily on prior knowledge, RL can learn directly from the feedback. On the other side, without a careful model design, existin… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
90
0
2

Year Published

2019
2019
2023
2023

Publication Types

Select...
4
2
1

Relationship

2
5

Authors

Journals

citations
Cited by 156 publications
(92 citation statements)
references
References 26 publications
0
90
0
2
Order By: Relevance
“…They also design a phase gate combined with a memory palace in the Q-network to distinguish the decision process for different phases and prevent the decision from favoring certain actions. FRAP [12] proposes a modified network structure to capture the phase competition relation between different traffic movements and speed up the training process. CoLight [15], the current stateof-the-art TSC method in multi-intersection scenario, utilizes graph attention on observations between agents to achieve cooperation.…”
Section: Related Workmentioning
confidence: 99%
See 3 more Smart Citations
“…They also design a phase gate combined with a memory palace in the Q-network to distinguish the decision process for different phases and prevent the decision from favoring certain actions. FRAP [12] proposes a modified network structure to capture the phase competition relation between different traffic movements and speed up the training process. CoLight [15], the current stateof-the-art TSC method in multi-intersection scenario, utilizes graph attention on observations between agents to achieve cooperation.…”
Section: Related Workmentioning
confidence: 99%
“…To estimate the reward function, some methods compute the weighted sum of several measurements of traffic situations (e.g., the number of vehicles, delay, etc.) [1], [9], [20] as reward while LIT [37] proves that using the queue length as reward is equivalent to minimizing average and some works follow this idea [12], [15]. However, this statement only holds in single intersection scenario.…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations
“…Note that state transition matrix P is not described in model-free methods. For traffic signal control, our RL agent is defined the same as [15].…”
Section: Background 31 Reinforcement Learningmentioning
confidence: 99%