Loading…

Traffic signal optimization through discrete and continuous reinforcement learning with robustness analysis in downtown Tehran

Traffic signal control plays a pivotal role in reducing traffic congestion. Traffic signals cannot be adequately controlled with conventional methods due to the high variations and complexity in traffic environments. In recent years, reinforcement learning (RL) has shown great potential for traffic...

Full description

Saved in:
Bibliographic Details
Published in:Advanced engineering informatics 2018-10, Vol.38, p.639-655
Main Authors: Aslani, Mohammad, Seipel, Stefan, Mesgari, Mohammad Saadi, Wiering, Marco
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Items that cite this one
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by cdi_FETCH-LOGICAL-c414t-f3379f53fd36bf37b6e152ca80a40646b7bf01af73d1060a15dcd98aa112a523
cites cdi_FETCH-LOGICAL-c414t-f3379f53fd36bf37b6e152ca80a40646b7bf01af73d1060a15dcd98aa112a523
container_end_page 655
container_issue
container_start_page 639
container_title Advanced engineering informatics
container_volume 38
creator Aslani, Mohammad
Seipel, Stefan
Mesgari, Mohammad Saadi
Wiering, Marco
description Traffic signal control plays a pivotal role in reducing traffic congestion. Traffic signals cannot be adequately controlled with conventional methods due to the high variations and complexity in traffic environments. In recent years, reinforcement learning (RL) has shown great potential for traffic signal control because of its high adaptability, flexibility, and scalability. However, designing RL-embedded traffic signal controllers (RLTSCs) for traffic systems with a high degree of realism is faced with several challenges, among others system disturbances and large state-action spaces are considered in this research. The contribution of the present work is founded on three features: (a) evaluating the robustness of different RLTSCs against system disturbances including incidents, jaywalking, and sensor noise, (b) handling a high-dimensional state-action space by both employing different continuous state RL algorithms and reducing the state-action space in order to improve the performance and learning speed of the system, and (c) presenting a detailed empirical study of traffic signals control of downtown Tehran through seven RL algorithms: discrete state Q-learning(λ), SARSA(λ), actor-critic(λ), continuous state Q-learning(λ), SARSA(λ), actor-critic(λ), and residual actor-critic(λ). In this research, first a real-world microscopic traffic simulation of downtown Tehran is carried out, then four experiments are performed in order to find the best RLTSC with convincing robustness and strong performance. The results reveal that the RLTSC based on continuous state actor-critic(λ) has the best performance. In addition, it is found that the best RLTSC leads to saving average travel time by 22% (at the presence of high system disturbances) when it is compared with an optimized fixed-time controller.
doi_str_mv 10.1016/j.aei.2018.08.002
format article
fullrecord <record><control><sourceid>elsevier_swepu</sourceid><recordid>TN_cdi_swepub_primary_oai_DiVA_org_uu_373216</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S1474034617302598</els_id><sourcerecordid>S1474034617302598</sourcerecordid><originalsourceid>FETCH-LOGICAL-c414t-f3379f53fd36bf37b6e152ca80a40646b7bf01af73d1060a15dcd98aa112a523</originalsourceid><addsrcrecordid>eNqNkc9qHDEMxk1poem2D9CbHyCztcfzb-kppEkbCPSy9Go0tjyjZddebE-X5NBnr5cNPYaKT0gH_YTEx9hnKdZSyO7Lbg1I61rIYS2KRP2GXcmhV1WravG29E3fVEI13Xv2IaWdKMyw6a_Yn20E58jwRJOHPQ_HTAd6hkzB8zzHsEwzt5RMxIwcvOUm-Ex-CUviEcm7EA0e0Ge-R4ie_MRPlGcew7ik7DGlQsH-KVHi5LkNJ59L8i3OEfxH9s7BPuGnl7pi2_u77e2P6vHn94fbm8fKNLLJlVOq37hWOau60al-7FC2tYFBQCO6phv70QkJrldWik6AbK2xmwFAyhraWq3Y9WVtOuFxGfUx0gHikw5A-hv9utEhTnpZtOpVLbv_G59p0vWgSqyYvIybGFKK6P4BUuizO3qnizv67I4WReJ80dcLg-Xr34RRJ0PoDVqKaLK2gV6h_wKufZxm</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Traffic signal optimization through discrete and continuous reinforcement learning with robustness analysis in downtown Tehran</title><source>Elsevier</source><creator>Aslani, Mohammad ; Seipel, Stefan ; Mesgari, Mohammad Saadi ; Wiering, Marco</creator><creatorcontrib>Aslani, Mohammad ; Seipel, Stefan ; Mesgari, Mohammad Saadi ; Wiering, Marco</creatorcontrib><description>Traffic signal control plays a pivotal role in reducing traffic congestion. Traffic signals cannot be adequately controlled with conventional methods due to the high variations and complexity in traffic environments. In recent years, reinforcement learning (RL) has shown great potential for traffic signal control because of its high adaptability, flexibility, and scalability. However, designing RL-embedded traffic signal controllers (RLTSCs) for traffic systems with a high degree of realism is faced with several challenges, among others system disturbances and large state-action spaces are considered in this research. The contribution of the present work is founded on three features: (a) evaluating the robustness of different RLTSCs against system disturbances including incidents, jaywalking, and sensor noise, (b) handling a high-dimensional state-action space by both employing different continuous state RL algorithms and reducing the state-action space in order to improve the performance and learning speed of the system, and (c) presenting a detailed empirical study of traffic signals control of downtown Tehran through seven RL algorithms: discrete state Q-learning(λ), SARSA(λ), actor-critic(λ), continuous state Q-learning(λ), SARSA(λ), actor-critic(λ), and residual actor-critic(λ). In this research, first a real-world microscopic traffic simulation of downtown Tehran is carried out, then four experiments are performed in order to find the best RLTSC with convincing robustness and strong performance. The results reveal that the RLTSC based on continuous state actor-critic(λ) has the best performance. In addition, it is found that the best RLTSC leads to saving average travel time by 22% (at the presence of high system disturbances) when it is compared with an optimized fixed-time controller.</description><identifier>ISSN: 1474-0346</identifier><identifier>ISSN: 1873-5320</identifier><identifier>EISSN: 1873-5320</identifier><identifier>DOI: 10.1016/j.aei.2018.08.002</identifier><language>eng</language><publisher>Elsevier Ltd</publisher><subject>Hållbar stadsutveckling ; Microscopic traffic simulation ; Reinforcement learning ; Sustainable Urban Development ; System disturbances ; Traffic signal control</subject><ispartof>Advanced engineering informatics, 2018-10, Vol.38, p.639-655</ispartof><rights>2018 Elsevier Ltd</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c414t-f3379f53fd36bf37b6e152ca80a40646b7bf01af73d1060a15dcd98aa112a523</citedby><cites>FETCH-LOGICAL-c414t-f3379f53fd36bf37b6e152ca80a40646b7bf01af73d1060a15dcd98aa112a523</cites><orcidid>0000-0003-0085-5829</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>230,314,780,784,885,27924,27925</link.rule.ids><backlink>$$Uhttps://urn.kb.se/resolve?urn=urn:nbn:se:hig:diva-28333$$DView record from Swedish Publication Index$$Hfree_for_read</backlink><backlink>$$Uhttps://urn.kb.se/resolve?urn=urn:nbn:se:uu:diva-373216$$DView record from Swedish Publication Index$$Hfree_for_read</backlink></links><search><creatorcontrib>Aslani, Mohammad</creatorcontrib><creatorcontrib>Seipel, Stefan</creatorcontrib><creatorcontrib>Mesgari, Mohammad Saadi</creatorcontrib><creatorcontrib>Wiering, Marco</creatorcontrib><title>Traffic signal optimization through discrete and continuous reinforcement learning with robustness analysis in downtown Tehran</title><title>Advanced engineering informatics</title><description>Traffic signal control plays a pivotal role in reducing traffic congestion. Traffic signals cannot be adequately controlled with conventional methods due to the high variations and complexity in traffic environments. In recent years, reinforcement learning (RL) has shown great potential for traffic signal control because of its high adaptability, flexibility, and scalability. However, designing RL-embedded traffic signal controllers (RLTSCs) for traffic systems with a high degree of realism is faced with several challenges, among others system disturbances and large state-action spaces are considered in this research. The contribution of the present work is founded on three features: (a) evaluating the robustness of different RLTSCs against system disturbances including incidents, jaywalking, and sensor noise, (b) handling a high-dimensional state-action space by both employing different continuous state RL algorithms and reducing the state-action space in order to improve the performance and learning speed of the system, and (c) presenting a detailed empirical study of traffic signals control of downtown Tehran through seven RL algorithms: discrete state Q-learning(λ), SARSA(λ), actor-critic(λ), continuous state Q-learning(λ), SARSA(λ), actor-critic(λ), and residual actor-critic(λ). In this research, first a real-world microscopic traffic simulation of downtown Tehran is carried out, then four experiments are performed in order to find the best RLTSC with convincing robustness and strong performance. The results reveal that the RLTSC based on continuous state actor-critic(λ) has the best performance. In addition, it is found that the best RLTSC leads to saving average travel time by 22% (at the presence of high system disturbances) when it is compared with an optimized fixed-time controller.</description><subject>Hållbar stadsutveckling</subject><subject>Microscopic traffic simulation</subject><subject>Reinforcement learning</subject><subject>Sustainable Urban Development</subject><subject>System disturbances</subject><subject>Traffic signal control</subject><issn>1474-0346</issn><issn>1873-5320</issn><issn>1873-5320</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2018</creationdate><recordtype>article</recordtype><recordid>eNqNkc9qHDEMxk1poem2D9CbHyCztcfzb-kppEkbCPSy9Go0tjyjZddebE-X5NBnr5cNPYaKT0gH_YTEx9hnKdZSyO7Lbg1I61rIYS2KRP2GXcmhV1WravG29E3fVEI13Xv2IaWdKMyw6a_Yn20E58jwRJOHPQ_HTAd6hkzB8zzHsEwzt5RMxIwcvOUm-Ex-CUviEcm7EA0e0Ge-R4ie_MRPlGcew7ik7DGlQsH-KVHi5LkNJ59L8i3OEfxH9s7BPuGnl7pi2_u77e2P6vHn94fbm8fKNLLJlVOq37hWOau60al-7FC2tYFBQCO6phv70QkJrldWik6AbK2xmwFAyhraWq3Y9WVtOuFxGfUx0gHikw5A-hv9utEhTnpZtOpVLbv_G59p0vWgSqyYvIybGFKK6P4BUuizO3qnizv67I4WReJ80dcLg-Xr34RRJ0PoDVqKaLK2gV6h_wKufZxm</recordid><startdate>20181001</startdate><enddate>20181001</enddate><creator>Aslani, Mohammad</creator><creator>Seipel, Stefan</creator><creator>Mesgari, Mohammad Saadi</creator><creator>Wiering, Marco</creator><general>Elsevier Ltd</general><scope>AAYXX</scope><scope>CITATION</scope><scope>ADTPV</scope><scope>AOWAS</scope><scope>D8W</scope><scope>DF2</scope><orcidid>https://orcid.org/0000-0003-0085-5829</orcidid></search><sort><creationdate>20181001</creationdate><title>Traffic signal optimization through discrete and continuous reinforcement learning with robustness analysis in downtown Tehran</title><author>Aslani, Mohammad ; Seipel, Stefan ; Mesgari, Mohammad Saadi ; Wiering, Marco</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c414t-f3379f53fd36bf37b6e152ca80a40646b7bf01af73d1060a15dcd98aa112a523</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2018</creationdate><topic>Hållbar stadsutveckling</topic><topic>Microscopic traffic simulation</topic><topic>Reinforcement learning</topic><topic>Sustainable Urban Development</topic><topic>System disturbances</topic><topic>Traffic signal control</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Aslani, Mohammad</creatorcontrib><creatorcontrib>Seipel, Stefan</creatorcontrib><creatorcontrib>Mesgari, Mohammad Saadi</creatorcontrib><creatorcontrib>Wiering, Marco</creatorcontrib><collection>CrossRef</collection><collection>SwePub</collection><collection>SwePub Articles</collection><collection>SWEPUB Högskolan i Gävle</collection><collection>SWEPUB Uppsala universitet</collection><jtitle>Advanced engineering informatics</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Aslani, Mohammad</au><au>Seipel, Stefan</au><au>Mesgari, Mohammad Saadi</au><au>Wiering, Marco</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Traffic signal optimization through discrete and continuous reinforcement learning with robustness analysis in downtown Tehran</atitle><jtitle>Advanced engineering informatics</jtitle><date>2018-10-01</date><risdate>2018</risdate><volume>38</volume><spage>639</spage><epage>655</epage><pages>639-655</pages><issn>1474-0346</issn><issn>1873-5320</issn><eissn>1873-5320</eissn><abstract>Traffic signal control plays a pivotal role in reducing traffic congestion. Traffic signals cannot be adequately controlled with conventional methods due to the high variations and complexity in traffic environments. In recent years, reinforcement learning (RL) has shown great potential for traffic signal control because of its high adaptability, flexibility, and scalability. However, designing RL-embedded traffic signal controllers (RLTSCs) for traffic systems with a high degree of realism is faced with several challenges, among others system disturbances and large state-action spaces are considered in this research. The contribution of the present work is founded on three features: (a) evaluating the robustness of different RLTSCs against system disturbances including incidents, jaywalking, and sensor noise, (b) handling a high-dimensional state-action space by both employing different continuous state RL algorithms and reducing the state-action space in order to improve the performance and learning speed of the system, and (c) presenting a detailed empirical study of traffic signals control of downtown Tehran through seven RL algorithms: discrete state Q-learning(λ), SARSA(λ), actor-critic(λ), continuous state Q-learning(λ), SARSA(λ), actor-critic(λ), and residual actor-critic(λ). In this research, first a real-world microscopic traffic simulation of downtown Tehran is carried out, then four experiments are performed in order to find the best RLTSC with convincing robustness and strong performance. The results reveal that the RLTSC based on continuous state actor-critic(λ) has the best performance. In addition, it is found that the best RLTSC leads to saving average travel time by 22% (at the presence of high system disturbances) when it is compared with an optimized fixed-time controller.</abstract><pub>Elsevier Ltd</pub><doi>10.1016/j.aei.2018.08.002</doi><tpages>17</tpages><orcidid>https://orcid.org/0000-0003-0085-5829</orcidid><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 1474-0346
ispartof Advanced engineering informatics, 2018-10, Vol.38, p.639-655
issn 1474-0346
1873-5320
1873-5320
language eng
recordid cdi_swepub_primary_oai_DiVA_org_uu_373216
source Elsevier
subjects Hållbar stadsutveckling
Microscopic traffic simulation
Reinforcement learning
Sustainable Urban Development
System disturbances
Traffic signal control
title Traffic signal optimization through discrete and continuous reinforcement learning with robustness analysis in downtown Tehran
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-27T08%3A18%3A12IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-elsevier_swepu&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Traffic%20signal%20optimization%20through%20discrete%20and%20continuous%20reinforcement%20learning%20with%20robustness%20analysis%20in%20downtown%20Tehran&rft.jtitle=Advanced%20engineering%20informatics&rft.au=Aslani,%20Mohammad&rft.date=2018-10-01&rft.volume=38&rft.spage=639&rft.epage=655&rft.pages=639-655&rft.issn=1474-0346&rft.eissn=1873-5320&rft_id=info:doi/10.1016/j.aei.2018.08.002&rft_dat=%3Celsevier_swepu%3ES1474034617302598%3C/elsevier_swepu%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c414t-f3379f53fd36bf37b6e152ca80a40646b7bf01af73d1060a15dcd98aa112a523%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true