Your browser doesn't support javascript.
loading
Deep Reinforcement Learning for Traffic Signal Control Model and Adaptation Study.
Tan, Jiyuan; Yuan, Qian; Guo, Weiwei; Xie, Na; Liu, Fuyu; Wei, Jing; Zhang, Xinwei.
Afiliación
  • Tan J; School of Electrical and Control Engineering, North China University of Technology, Beijing 100144, China.
  • Yuan Q; School of Electrical and Control Engineering, North China University of Technology, Beijing 100144, China.
  • Guo W; School of Electrical and Control Engineering, North China University of Technology, Beijing 100144, China.
  • Xie N; School of Management Science and Engineering, Central University of Finance and Economics, Beijing 100081, China.
  • Liu F; School of Electrical and Control Engineering, North China University of Technology, Beijing 100144, China.
  • Wei J; School of Electrical and Control Engineering, North China University of Technology, Beijing 100144, China.
  • Zhang X; School of Electrical and Control Engineering, North China University of Technology, Beijing 100144, China.
Sensors (Basel) ; 22(22)2022 Nov 11.
Article en En | MEDLINE | ID: mdl-36433328
ABSTRACT
Deep reinforcement learning provides a new approach to solving complex signal optimization problems at intersections. Earlier studies were limited to traditional traffic detection techniques, and the obtained traffic information was not accurate. With the advanced in technology, we can obtain highly accurate information on the traffic states by advanced detector technology. This provides an accurate source of data for deep reinforcement learning. There are many intersections in the urban network. To successfully apply deep reinforcement learning in a situation closer to reality, we need to consider the problem of extending the knowledge gained from the training to new scenarios. This study used advanced sensor technology as a data source to explore the variation pattern of state space under different traffic scenarios. It analyzes the relationship between the traffic demand and the actual traffic states. The model learned more from a more comprehensive state space of traffic. This model was successful applied to new traffic scenarios without additional training. Compared our proposed model with the popular SAC signal control model, the result shows that the average delay of the DQN model is 5.13 s and the SAC model is 6.52 s. Therefore, our model exhibits better control performance.
Asunto(s)
Palabras clave

Texto completo: 1 Colección: 01-internacional Base de datos: MEDLINE Asunto principal: Refuerzo en Psicología / Aclimatación Idioma: En Revista: Sensors (Basel) Año: 2022 Tipo del documento: Article País de afiliación: China

Texto completo: 1 Colección: 01-internacional Base de datos: MEDLINE Asunto principal: Refuerzo en Psicología / Aclimatación Idioma: En Revista: Sensors (Basel) Año: 2022 Tipo del documento: Article País de afiliación: China