Skip to main content
Erschienen in: International Journal of Intelligent Transportation Systems Research 1/2023

02.02.2023

Deep Reinforcement Q-Learning for Intelligent Traffic Signal Control with Partial Detection

verfasst von: Romain Ducrocq, Nadir Farhi

Erschienen in: International Journal of Intelligent Transportation Systems Research | Ausgabe 1/2023

Einloggen

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Intelligent traffic signal controllers, applying DQN algorithms to traffic light policy optimization, efficiently reduce traffic congestion by adjusting traffic signals to real-time traffic. Most propositions in the literature however consider that all vehicles at the intersection are detected, an unrealistic scenario. Recently, new wireless communication technologies have enabled cost-efficient detection of connected vehicles by infrastructures. With only a small fraction of the total fleet currently equipped, methods able to perform under low detection rates are desirable. In this paper, we propose a deep reinforcement Q-learning model to optimize traffic signal control at an isolated intersection, in a partially observable environment with connected vehicles. First, we present the novel DQN model within the RL framework. We introduce a new state representation for partially observable environments and a new reward function for traffic signal control, and provide a network architecture and tuned hyper-parameters. Second, we evaluate the performances of the model in numerical simulations on multiple scenarios, in two steps. At first in full detection against existing actuated controllers, then in partial detection with loss estimates for proportions of connected vehicles. Finally, from the obtained results, we define thresholds for detection rates with acceptable and optimal performance levels. The source code implementation of the model is available at: https://​github.​com/​romainducrocq/​DQN-ITSCwPD

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

ATZelectronics worldwide

ATZlectronics worldwide is up-to-speed on new trends and developments in automotive electronics on a scientific level with a high depth of information. 

Order your 30-days-trial for free and without any commitment.

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Weitere Produktempfehlungen anzeigen
Fußnoten
1
Fairness is to be understood here as follows: A distribution d1 of delays over vehicles is fairer than another distribution d2 of delays over vehicles, if d1 is closer to the uniform distribution, compared to d2.
 
Literatur
1.
Zurück zum Zitat Alam, J., ey, M.K.: Design and analysis of a two stage traffic light system using fuzzy logic. J. Inf. Technol. Softw. Eng. 5, 1–9 (2015) Alam, J., ey, M.K.: Design and analysis of a two stage traffic light system using fuzzy logic. J. Inf. Technol. Softw. Eng. 5, 1–9 (2015)
2.
Zurück zum Zitat Alemzadeh, S., Moslemi, R., Sharma, R., Mesbahi, M.: Adaptive traffic control with deep reinforcement learning towards state-of-the-art and beyond. arXiv:2007.10960 (2020) Alemzadeh, S., Moslemi, R., Sharma, R., Mesbahi, M.: Adaptive traffic control with deep reinforcement learning towards state-of-the-art and beyond. arXiv:2007.​10960 (2020)
4.
Zurück zum Zitat Codecà, L., Frank, R., Engel, T.: Luxembourg sumo traffic (lust) scenario: 24 hours of mobility for vehicular networking research. In: 2015 IEEE vehicular networking conference (VNC), pp. 1–8 (2015) Codecà, L., Frank, R., Engel, T.: Luxembourg sumo traffic (lust) scenario: 24 hours of mobility for vehicular networking research. In: 2015 IEEE vehicular networking conference (VNC), pp. 1–8 (2015)
5.
Zurück zum Zitat Gao, J., Shen, Y., Liu, J., Ito, M., Shiratori, N.: Adaptive traffic signal control deep reinforcement learning algorithm with experience replay and target network. arXiv:1705.02755 (2017) Gao, J., Shen, Y., Liu, J., Ito, M., Shiratori, N.: Adaptive traffic signal control deep reinforcement learning algorithm with experience replay and target network. arXiv:1705.​02755 (2017)
6.
Zurück zum Zitat Genders, W., Razavi, S.N.: Evaluating reinforcement learning state representations for adaptive traffic signal control. In: ANT/SEIT (2018) Genders, W., Razavi, S.N.: Evaluating reinforcement learning state representations for adaptive traffic signal control. In: ANT/SEIT (2018)
7.
Zurück zum Zitat Genders, W., Razavi, S.N.: An open-source framework for adaptive traffic signal control. arXiv:1909.00395 (2019) Genders, W., Razavi, S.N.: An open-source framework for adaptive traffic signal control. arXiv:1909.​00395 (2019)
8.
Zurück zum Zitat Gershenson, C.: Self-organizing traffic lights. Complex Syst.:16 (2005) Gershenson, C.: Self-organizing traffic lights. Complex Syst.:16 (2005)
9.
Zurück zum Zitat Hasselt, H.V., Guez, A., Silver, D.: Deep reinforcement learning with double q-learning. arXiv:1509.06461 (2016) Hasselt, H.V., Guez, A., Silver, D.: Deep reinforcement learning with double q-learning. arXiv:1509.​06461 (2016)
10.
Zurück zum Zitat Hessel, M., Modayil, J., Hasselt, H.V., Schaul, T., Ostrovski, G., Dabney, W., Horgan, D., Piot, B., Azar, M.G., Silver, D.: Rainbow: Combining improvements in deep reinforcement learning. In: AAAI (2018) Hessel, M., Modayil, J., Hasselt, H.V., Schaul, T., Ostrovski, G., Dabney, W., Horgan, D., Piot, B., Azar, M.G., Silver, D.: Rainbow: Combining improvements in deep reinforcement learning. In: AAAI (2018)
11.
Zurück zum Zitat Kheterpal, N., Parvate, K., Wu, C., Kreidieh, A., Vinitsky, E., Bayen, A.M.: Flow: deep reinforcement learning for control in sumo (2018) Kheterpal, N., Parvate, K., Wu, C., Kreidieh, A., Vinitsky, E., Bayen, A.M.: Flow: deep reinforcement learning for control in sumo (2018)
12.
Zurück zum Zitat Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. CoRR arXiv:1412.6980(2015) Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. CoRR arXiv:1412.​6980(2015)
13.
Zurück zum Zitat Krajzewicz, D., Erdmann, J., Behrisch, M., Bieker, L.: Recent development and applications of sumo – simulation of urban mobility (2012) Krajzewicz, D., Erdmann, J., Behrisch, M., Bieker, L.: Recent development and applications of sumo – simulation of urban mobility (2012)
14.
Zurück zum Zitat Krizhevsky, A., Sutskever, I., Hinton, G.E.: Imagenet classification with deep convolutional neural networks. Commun. ACM 60, 84–90 (2012)CrossRef Krizhevsky, A., Sutskever, I., Hinton, G.E.: Imagenet classification with deep convolutional neural networks. Commun. ACM 60, 84–90 (2012)CrossRef
15.
Zurück zum Zitat Li, L., Lv, Y., Wang, F.: Traffic signal timing via deep reinforcement learning. IEEE/CAA J. Autom. Sinica 3, 247–254 (2016)MathSciNetCrossRef Li, L., Lv, Y., Wang, F.: Traffic signal timing via deep reinforcement learning. IEEE/CAA J. Autom. Sinica 3, 247–254 (2016)MathSciNetCrossRef
16.
Zurück zum Zitat Liang, X., Du, X., Wang, G., Han, Z.: A deep reinforcement learning network for traffic light cycle control. IEEE Trans. Veh. Technol. 68, 1243–1253 (2019)CrossRef Liang, X., Du, X., Wang, G., Han, Z.: A deep reinforcement learning network for traffic light cycle control. IEEE Trans. Veh. Technol. 68, 1243–1253 (2019)CrossRef
17.
Zurück zum Zitat Lillicrap, T.P., Hunt, J.J., Pritzel, A., Heess, N.M.O., Erez, T., Tassa, Y., Silver, D., Wierstra, D.: Continuous control with deep reinforcement learning. CoRR arXiv:1509.02971 (2016) Lillicrap, T.P., Hunt, J.J., Pritzel, A., Heess, N.M.O., Erez, T., Tassa, Y., Silver, D., Wierstra, D.: Continuous control with deep reinforcement learning. CoRR arXiv:1509.​02971 (2016)
18.
Zurück zum Zitat Mnih, V., Kavukcuoglu, K., Silver, D., Graves, A., Antonoglou, I., Wierstra, D., Riedmiller, M.A.: Playing atari with deep reinforcement learning. arXiv:1312.5602 (2013) Mnih, V., Kavukcuoglu, K., Silver, D., Graves, A., Antonoglou, I., Wierstra, D., Riedmiller, M.A.: Playing atari with deep reinforcement learning. arXiv:1312.​5602 (2013)
19.
Zurück zum Zitat Mnih, V., Kavukcuoglu, K., Silver, D., Rusu, A.A., Veness, J., Bellemare, M.G., Graves, A., Riedmiller, M.A., Fidjeland, A., Ostrovski, G., Petersen, S., Beattie, C., Sadik, A., Antonoglou, I., King, H., Kumaran, D., Wierstra, D., Legg, S., Hassabis, D.: Human-level control through deep reinforcement learning. Nature 518, 529–533 (2015)CrossRef Mnih, V., Kavukcuoglu, K., Silver, D., Rusu, A.A., Veness, J., Bellemare, M.G., Graves, A., Riedmiller, M.A., Fidjeland, A., Ostrovski, G., Petersen, S., Beattie, C., Sadik, A., Antonoglou, I., King, H., Kumaran, D., Wierstra, D., Legg, S., Hassabis, D.: Human-level control through deep reinforcement learning. Nature 518, 529–533 (2015)CrossRef
20.
Zurück zum Zitat Nguyen Van Phu, C., Farhi, N.: Estimation of urban traffic state with probe vehicles. IEEE Trans. Intell. Transp. Syst. 22, 2797–2808 (2021)CrossRef Nguyen Van Phu, C., Farhi, N.: Estimation of urban traffic state with probe vehicles. IEEE Trans. Intell. Transp. Syst. 22, 2797–2808 (2021)CrossRef
21.
Zurück zum Zitat Płaczek, B.: A self-organizing system for urban traffic control based on predictive interval microscopic model. arXiv:1406.1128 (2014) Płaczek, B.: A self-organizing system for urban traffic control based on predictive interval microscopic model. arXiv:1406.​1128 (2014)
22.
Zurück zum Zitat Stevens, M., Yeh, C.: Reinforcement learning for traffic optimization (2016) Stevens, M., Yeh, C.: Reinforcement learning for traffic optimization (2016)
24.
Zurück zum Zitat Touhbi, S., Babram, M.A., Nguyen-Huu, T., Marilleau, N., Hbid, M.L., Cambier, C., Stinckwich, S.: Adaptive traffic signal control : exploring reward definition for reinforcement learning. In: ANT/SEIT (2017) Touhbi, S., Babram, M.A., Nguyen-Huu, T., Marilleau, N., Hbid, M.L., Cambier, C., Stinckwich, S.: Adaptive traffic signal control : exploring reward definition for reinforcement learning. In: ANT/SEIT (2017)
25.
Zurück zum Zitat van der Pol, E., Oliehoek, F.A.: Coordinated deep reinforcement learners for traffic light control (2016) van der Pol, E., Oliehoek, F.A.: Coordinated deep reinforcement learners for traffic light control (2016)
26.
Zurück zum Zitat Varaiya, P.: Max pressure control of a network of signalized intersections. Trans. Res. Part C-emerging Technol. 36, 177–195 (2013)CrossRef Varaiya, P.: Max pressure control of a network of signalized intersections. Trans. Res. Part C-emerging Technol. 36, 177–195 (2013)CrossRef
27.
Zurück zum Zitat Vidali, A., Crociani, L., Vizzari, G., Bandini, S.: A deep reinforcement learning approach to adaptive traffic lights management. In: WOA (2019) Vidali, A., Crociani, L., Vizzari, G., Bandini, S.: A deep reinforcement learning approach to adaptive traffic lights management. In: WOA (2019)
28.
Zurück zum Zitat Wang, Z., Schaul, T., Hessel, M., Hasselt, H.V., Lanctot, M., de Freitas, N.: Dueling network architectures for deep reinforcement learning. arXiv:1511.06581 (2016) Wang, Z., Schaul, T., Hessel, M., Hasselt, H.V., Lanctot, M., de Freitas, N.: Dueling network architectures for deep reinforcement learning. arXiv:1511.​06581 (2016)
30.
Zurück zum Zitat Wei, H., Chen, C., Zheng, G., Wu, K., Gayah, V.V., Xu, K., Li, Z.J.: Presslight: learning max pressure control to coordinate traffic signals in arterial network. In: Proceedings of the 25th ACM SIGKDD international conference on knowledge discovery & data mining (2019a) Wei, H., Chen, C., Zheng, G., Wu, K., Gayah, V.V., Xu, K., Li, Z.J.: Presslight: learning max pressure control to coordinate traffic signals in arterial network. In: Proceedings of the 25th ACM SIGKDD international conference on knowledge discovery & data mining (2019a)
31.
Zurück zum Zitat Wei, H., Xu, N., Zhang, H., Zheng, G., Zang, X., Chen, C., Zhang, W., Zhu, Y., Xu, K., Li, Z.: Colight: Learning network-level cooperation for traffic signal control. In: Proceedings of the 28th ACM international conference on information and knowledge management, CIKM ’19, pp. 1913–1922. New York. Association for Computing Machinery. ISBN 9781450369763. https://doi.org/10.1145/3357384.3357902 (2019b) Wei, H., Xu, N., Zhang, H., Zheng, G., Zang, X., Chen, C., Zhang, W., Zhu, Y., Xu, K., Li, Z.: Colight: Learning network-level cooperation for traffic signal control. In: Proceedings of the 28th ACM international conference on information and knowledge management, CIKM ’19, pp. 1913–1922. New York. Association for Computing Machinery. ISBN 9781450369763. https://​doi.​org/​10.​1145/​3357384.​3357902 (2019b)
32.
Zurück zum Zitat Wei, H., Zheng, G., Gayah, V.V., Li, Z.J.: A survey on traffic signal control methods. arXiv:1904.08117 (2019c) Wei, H., Zheng, G., Gayah, V.V., Li, Z.J.: A survey on traffic signal control methods. arXiv:1904.​08117 (2019c)
33.
Zurück zum Zitat Yan, S., Zhang, J., Büscher, D., Burgard, W.: Efficiency and equity are both essential: A generalized traffic signal controller with deep reinforcement learning. In: 2020 IEEE/RSJ international conference on intelligent robots and systems (IROS), pp. 5526–5533. https://doi.org/10.1109/IROS45743.2020.9340784 (2020) Yan, S., Zhang, J., Büscher, D., Burgard, W.: Efficiency and equity are both essential: A generalized traffic signal controller with deep reinforcement learning. In: 2020 IEEE/RSJ international conference on intelligent robots and systems (IROS), pp. 5526–5533. https://​doi.​org/​10.​1109/​IROS45743.​2020.​9340784 (2020)
34.
Zurück zum Zitat Zhang, R., Leteurtre, R., Striner, B., Alanazi, A.S., Alghafis, A., Tonguz, O.K.: Partially detected intelligent traffic signal control: environmental adaptation. In: 2019 18th IEEE international conference on machine learning and applications (ICMLA), pp. 1956–1960 (2019) Zhang, R., Leteurtre, R., Striner, B., Alanazi, A.S., Alghafis, A., Tonguz, O.K.: Partially detected intelligent traffic signal control: environmental adaptation. In: 2019 18th IEEE international conference on machine learning and applications (ICMLA), pp. 1956–1960 (2019)
35.
Zurück zum Zitat Zhang, R., Ishikawa, A., Wang, W., Striner, B., Tonguz, O.K.: Using reinforcement learning with partial vehicle detection for intelligent traffic signal control. IEEE Trans. Intell. Transp. Syst. 22, 404–415 (2021)CrossRef Zhang, R., Ishikawa, A., Wang, W., Striner, B., Tonguz, O.K.: Using reinforcement learning with partial vehicle detection for intelligent traffic signal control. IEEE Trans. Intell. Transp. Syst. 22, 404–415 (2021)CrossRef
36.
Zurück zum Zitat Zheng, G., Xiong, Y., Zang, X., Feng, J., Wei, H., Zhang, H., Li, Y., Xu, K., Li, Z.: Learning phase competition for traffic signal control. In: Proceedings of the 28th ACM international conference on information and knowledge management, CIKM ’19, pp. 1963–1972. New York. Association for Computing Machinery. ISBN 9781450369763. https://doi.org/10.1145/3357384.3357900 (2019) Zheng, G., Xiong, Y., Zang, X., Feng, J., Wei, H., Zhang, H., Li, Y., Xu, K., Li, Z.: Learning phase competition for traffic signal control. In: Proceedings of the 28th ACM international conference on information and knowledge management, CIKM ’19, pp. 1963–1972. New York. Association for Computing Machinery. ISBN 9781450369763. https://​doi.​org/​10.​1145/​3357384.​3357900 (2019)
Metadaten
Titel
Deep Reinforcement Q-Learning for Intelligent Traffic Signal Control with Partial Detection
verfasst von
Romain Ducrocq
Nadir Farhi
Publikationsdatum
02.02.2023
Verlag
Springer US
Erschienen in
International Journal of Intelligent Transportation Systems Research / Ausgabe 1/2023
Print ISSN: 1348-8503
Elektronische ISSN: 1868-8659
DOI
https://doi.org/10.1007/s13177-023-00346-4

Weitere Artikel der Ausgabe 1/2023

International Journal of Intelligent Transportation Systems Research 1/2023 Zur Ausgabe

    Marktübersichten

    Die im Laufe eines Jahres in der „adhäsion“ veröffentlichten Marktübersichten helfen Anwendern verschiedenster Branchen, sich einen gezielten Überblick über Lieferantenangebote zu verschaffen.