Traffic Signal Control Using Hybrid Action Space Deep Reinforcement Learning

Recent research works on intelligent traffic signal control (TSC) have been mainly focused on leveraging deep reinforcement learning (DRL) due to its proven capability and performance. DRL-based traffic signal control frameworks belong to either discrete or continuous controls. In discrete control,...

Full description

Bibliographic Details
Main Authors: Salah Bouktif, Abderraouf Cheniki, Ali Ouni
Format: Article
Language:English
Published: MDPI AG 2021-03-01
Series:Sensors
Subjects:
Online Access:https://www.mdpi.com/1424-8220/21/7/2302
id doaj-59001efee2474721b816b4040517fbf1
record_format Article
spelling doaj-59001efee2474721b816b4040517fbf12021-03-26T00:05:19ZengMDPI AGSensors1424-82202021-03-01212302230210.3390/s21072302Traffic Signal Control Using Hybrid Action Space Deep Reinforcement LearningSalah Bouktif0Abderraouf Cheniki1Ali Ouni2Department of Computer Science and Software Engineering, University of United Arab Emirates, Al Ain 15551, Abu Dhabi, United Arab EmiratesDepartment of Electrical Engineering, University of Boumerdes, Boumerdès 35000, AlgeriaÉcole de Technologie Supérieure, University of Quebec, Montreal, QC H3C 1K3, CanadaRecent research works on intelligent traffic signal control (TSC) have been mainly focused on leveraging deep reinforcement learning (DRL) due to its proven capability and performance. DRL-based traffic signal control frameworks belong to either discrete or continuous controls. In discrete control, the DRL agent selects the appropriate traffic light phase from a finite set of phases. Whereas in continuous control approach, the agent decides the appropriate duration for each signal phase within a predetermined sequence of phases. Among the existing works, there are no prior approaches that propose a flexible framework combining both discrete and continuous DRL approaches in controlling traffic signal. Thus, our ultimate objective in this paper is to propose an approach capable of deciding simultaneously the proper phase and its associated duration. Our contribution resides in adapting a hybrid Deep Reinforcement Learning that considers at the same time discrete and continuous decisions. Precisely, we customize a Parameterized Deep Q-Networks (P-DQN) architecture that permits a hierarchical decision-making process that primarily decides the traffic light next phases and secondly specifies its the associated timing. The evaluation results of our approach using Simulation of Urban MObility (SUMO) shows its out-performance over the benchmarks. The proposed framework is able to reduce the average queue length of vehicles and the average travel time by 22.20% and 5.78%, respectively, over the alternative DRL-based TSC systems.https://www.mdpi.com/1424-8220/21/7/2302traffic signal controltraffic optimizationparameterized deep reinforcement learningP-DQNhybrid action space
collection DOAJ
language English
format Article
sources DOAJ
author Salah Bouktif
Abderraouf Cheniki
Ali Ouni
spellingShingle Salah Bouktif
Abderraouf Cheniki
Ali Ouni
Traffic Signal Control Using Hybrid Action Space Deep Reinforcement Learning
Sensors
traffic signal control
traffic optimization
parameterized deep reinforcement learning
P-DQN
hybrid action space
author_facet Salah Bouktif
Abderraouf Cheniki
Ali Ouni
author_sort Salah Bouktif
title Traffic Signal Control Using Hybrid Action Space Deep Reinforcement Learning
title_short Traffic Signal Control Using Hybrid Action Space Deep Reinforcement Learning
title_full Traffic Signal Control Using Hybrid Action Space Deep Reinforcement Learning
title_fullStr Traffic Signal Control Using Hybrid Action Space Deep Reinforcement Learning
title_full_unstemmed Traffic Signal Control Using Hybrid Action Space Deep Reinforcement Learning
title_sort traffic signal control using hybrid action space deep reinforcement learning
publisher MDPI AG
series Sensors
issn 1424-8220
publishDate 2021-03-01
description Recent research works on intelligent traffic signal control (TSC) have been mainly focused on leveraging deep reinforcement learning (DRL) due to its proven capability and performance. DRL-based traffic signal control frameworks belong to either discrete or continuous controls. In discrete control, the DRL agent selects the appropriate traffic light phase from a finite set of phases. Whereas in continuous control approach, the agent decides the appropriate duration for each signal phase within a predetermined sequence of phases. Among the existing works, there are no prior approaches that propose a flexible framework combining both discrete and continuous DRL approaches in controlling traffic signal. Thus, our ultimate objective in this paper is to propose an approach capable of deciding simultaneously the proper phase and its associated duration. Our contribution resides in adapting a hybrid Deep Reinforcement Learning that considers at the same time discrete and continuous decisions. Precisely, we customize a Parameterized Deep Q-Networks (P-DQN) architecture that permits a hierarchical decision-making process that primarily decides the traffic light next phases and secondly specifies its the associated timing. The evaluation results of our approach using Simulation of Urban MObility (SUMO) shows its out-performance over the benchmarks. The proposed framework is able to reduce the average queue length of vehicles and the average travel time by 22.20% and 5.78%, respectively, over the alternative DRL-based TSC systems.
topic traffic signal control
traffic optimization
parameterized deep reinforcement learning
P-DQN
hybrid action space
url https://www.mdpi.com/1424-8220/21/7/2302
work_keys_str_mv AT salahbouktif trafficsignalcontrolusinghybridactionspacedeepreinforcementlearning
AT abderraoufcheniki trafficsignalcontrolusinghybridactionspacedeepreinforcementlearning
AT aliouni trafficsignalcontrolusinghybridactionspacedeepreinforcementlearning
_version_ 1724203093630910464