Loading…

Dynamic holding control to avoid bus bunching: A multi-agent deep reinforcement learning framework

•A multi-agent deep reinforcement learning framework is proposed for bus holding control.•A reward function is defined to achieve headway self-equalization.•The action of each agent is considered by introducing a joint-action tracker.•A scheme based on proximal policy optimization is designed to tra...

Full description

Saved in:
Bibliographic Details
Published in:Transportation research. Part C, Emerging technologies Emerging technologies, 2020-07, Vol.116, p.102661, Article 102661
Main Authors: Wang, Jiawei, Sun, Lijun
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Items that cite this one
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by cdi_FETCH-LOGICAL-c325t-a9de5e2071c4b7fdbcc565523b0cb8787afc2e0e3acb51a71527ae8b472e6afc3
cites cdi_FETCH-LOGICAL-c325t-a9de5e2071c4b7fdbcc565523b0cb8787afc2e0e3acb51a71527ae8b472e6afc3
container_end_page
container_issue
container_start_page 102661
container_title Transportation research. Part C, Emerging technologies
container_volume 116
creator Wang, Jiawei
Sun, Lijun
description •A multi-agent deep reinforcement learning framework is proposed for bus holding control.•A reward function is defined to achieve headway self-equalization.•The action of each agent is considered by introducing a joint-action tracker.•A scheme based on proximal policy optimization is designed to train the agents.•The framework outperforms other baselines in simulation studies. Bus bunching has been a long-standing problem that undermines the efficiency and reliability of public transport services. The most popular countermeasure in practice is to introduce static and dynamic holding control. However, most previous holding control strategies mainly consider local information with a pre-specified headway/schedule, while the global coordination of the whole bus fleet and its long-term effect are often overlooked. To efficiently incorporate global coordination and long-term operation in bus holding, in this paper we propose a multi-agent deep reinforcement learning (MDRL) framework to develop dynamic and flexible holding control strategies for a bus route. Specifically, we model each bus as an agent that interacts with not only its leader/follower but also all other vehicles in the fleet. To better explore potential strategies, we develop an effective headway-based reward function in the proposed framework. In the learning framework, we model fleet coordination by using a basic actor-critic scheme along with a joint action tracker to better characterize the complex interactions among agents in policy learning, and we apply proximal policy optimization to improve learning performance. We conduct extensive numerical experiments to evaluate the proposed MDRL framework against multiple baseline models that only rely on local information. Our results demonstrate the superiority of the proposed framework and show the promise of applying MDRL in the coordinative control of public transport vehicle fleets in real-world operations.
doi_str_mv 10.1016/j.trc.2020.102661
format article
fullrecord <record><control><sourceid>elsevier_cross</sourceid><recordid>TN_cdi_crossref_primary_10_1016_j_trc_2020_102661</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S0968090X20305763</els_id><sourcerecordid>S0968090X20305763</sourcerecordid><originalsourceid>FETCH-LOGICAL-c325t-a9de5e2071c4b7fdbcc565523b0cb8787afc2e0e3acb51a71527ae8b472e6afc3</originalsourceid><addsrcrecordid>eNp9kM1OwzAQhC0EEqXwANz8Aim2U8cJnKryK1XiAhI3y95sWpfErhy3qG9PonLmsFrtjmY0-gi55WzGGS_utrMUYSaYGG9RFPyMTHipqkzksjonE1YVZcYq9nVJrvp-yxjjlVQTYh-P3nQO6Ca0tfNrCsGnGFqaAjWH4Gpq9_0wHjaDek8XtNu3yWVmjT7RGnFHIzrfhAjYja8WTfRjUBNNhz8hfl-Ti8a0Pd787Sn5fH76WL5mq_eXt-VilUEuZMpMVaNEwRSHuVVNbQFkIaXILQNbqlKZBgQyzA1YyY3iUiiDpZ0rgcWg5VPCT7kQQ99HbPQuus7Eo-ZMj5D0Vg-Q9AhJnyANnoeTB4diB4dR9-DQA9YuIiRdB_eP-xebLHIT</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Dynamic holding control to avoid bus bunching: A multi-agent deep reinforcement learning framework</title><source>ScienceDirect Freedom Collection</source><creator>Wang, Jiawei ; Sun, Lijun</creator><creatorcontrib>Wang, Jiawei ; Sun, Lijun</creatorcontrib><description>•A multi-agent deep reinforcement learning framework is proposed for bus holding control.•A reward function is defined to achieve headway self-equalization.•The action of each agent is considered by introducing a joint-action tracker.•A scheme based on proximal policy optimization is designed to train the agents.•The framework outperforms other baselines in simulation studies. Bus bunching has been a long-standing problem that undermines the efficiency and reliability of public transport services. The most popular countermeasure in practice is to introduce static and dynamic holding control. However, most previous holding control strategies mainly consider local information with a pre-specified headway/schedule, while the global coordination of the whole bus fleet and its long-term effect are often overlooked. To efficiently incorporate global coordination and long-term operation in bus holding, in this paper we propose a multi-agent deep reinforcement learning (MDRL) framework to develop dynamic and flexible holding control strategies for a bus route. Specifically, we model each bus as an agent that interacts with not only its leader/follower but also all other vehicles in the fleet. To better explore potential strategies, we develop an effective headway-based reward function in the proposed framework. In the learning framework, we model fleet coordination by using a basic actor-critic scheme along with a joint action tracker to better characterize the complex interactions among agents in policy learning, and we apply proximal policy optimization to improve learning performance. We conduct extensive numerical experiments to evaluate the proposed MDRL framework against multiple baseline models that only rely on local information. Our results demonstrate the superiority of the proposed framework and show the promise of applying MDRL in the coordinative control of public transport vehicle fleets in real-world operations.</description><identifier>ISSN: 0968-090X</identifier><identifier>EISSN: 1879-2359</identifier><identifier>DOI: 10.1016/j.trc.2020.102661</identifier><language>eng</language><publisher>Elsevier Ltd</publisher><subject>Bus bunching ; Deep reinforcement learning ; Dynamic holding ; Multi-agent system ; Policy proximal optimization</subject><ispartof>Transportation research. Part C, Emerging technologies, 2020-07, Vol.116, p.102661, Article 102661</ispartof><rights>2020 Elsevier Ltd</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c325t-a9de5e2071c4b7fdbcc565523b0cb8787afc2e0e3acb51a71527ae8b472e6afc3</citedby><cites>FETCH-LOGICAL-c325t-a9de5e2071c4b7fdbcc565523b0cb8787afc2e0e3acb51a71527ae8b472e6afc3</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,776,780,27901,27902</link.rule.ids></links><search><creatorcontrib>Wang, Jiawei</creatorcontrib><creatorcontrib>Sun, Lijun</creatorcontrib><title>Dynamic holding control to avoid bus bunching: A multi-agent deep reinforcement learning framework</title><title>Transportation research. Part C, Emerging technologies</title><description>•A multi-agent deep reinforcement learning framework is proposed for bus holding control.•A reward function is defined to achieve headway self-equalization.•The action of each agent is considered by introducing a joint-action tracker.•A scheme based on proximal policy optimization is designed to train the agents.•The framework outperforms other baselines in simulation studies. Bus bunching has been a long-standing problem that undermines the efficiency and reliability of public transport services. The most popular countermeasure in practice is to introduce static and dynamic holding control. However, most previous holding control strategies mainly consider local information with a pre-specified headway/schedule, while the global coordination of the whole bus fleet and its long-term effect are often overlooked. To efficiently incorporate global coordination and long-term operation in bus holding, in this paper we propose a multi-agent deep reinforcement learning (MDRL) framework to develop dynamic and flexible holding control strategies for a bus route. Specifically, we model each bus as an agent that interacts with not only its leader/follower but also all other vehicles in the fleet. To better explore potential strategies, we develop an effective headway-based reward function in the proposed framework. In the learning framework, we model fleet coordination by using a basic actor-critic scheme along with a joint action tracker to better characterize the complex interactions among agents in policy learning, and we apply proximal policy optimization to improve learning performance. We conduct extensive numerical experiments to evaluate the proposed MDRL framework against multiple baseline models that only rely on local information. Our results demonstrate the superiority of the proposed framework and show the promise of applying MDRL in the coordinative control of public transport vehicle fleets in real-world operations.</description><subject>Bus bunching</subject><subject>Deep reinforcement learning</subject><subject>Dynamic holding</subject><subject>Multi-agent system</subject><subject>Policy proximal optimization</subject><issn>0968-090X</issn><issn>1879-2359</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2020</creationdate><recordtype>article</recordtype><recordid>eNp9kM1OwzAQhC0EEqXwANz8Aim2U8cJnKryK1XiAhI3y95sWpfErhy3qG9PonLmsFrtjmY0-gi55WzGGS_utrMUYSaYGG9RFPyMTHipqkzksjonE1YVZcYq9nVJrvp-yxjjlVQTYh-P3nQO6Ca0tfNrCsGnGFqaAjWH4Gpq9_0wHjaDek8XtNu3yWVmjT7RGnFHIzrfhAjYja8WTfRjUBNNhz8hfl-Ti8a0Pd787Sn5fH76WL5mq_eXt-VilUEuZMpMVaNEwRSHuVVNbQFkIaXILQNbqlKZBgQyzA1YyY3iUiiDpZ0rgcWg5VPCT7kQQ99HbPQuus7Eo-ZMj5D0Vg-Q9AhJnyANnoeTB4diB4dR9-DQA9YuIiRdB_eP-xebLHIT</recordid><startdate>20200701</startdate><enddate>20200701</enddate><creator>Wang, Jiawei</creator><creator>Sun, Lijun</creator><general>Elsevier Ltd</general><scope>AAYXX</scope><scope>CITATION</scope></search><sort><creationdate>20200701</creationdate><title>Dynamic holding control to avoid bus bunching: A multi-agent deep reinforcement learning framework</title><author>Wang, Jiawei ; Sun, Lijun</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c325t-a9de5e2071c4b7fdbcc565523b0cb8787afc2e0e3acb51a71527ae8b472e6afc3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2020</creationdate><topic>Bus bunching</topic><topic>Deep reinforcement learning</topic><topic>Dynamic holding</topic><topic>Multi-agent system</topic><topic>Policy proximal optimization</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Wang, Jiawei</creatorcontrib><creatorcontrib>Sun, Lijun</creatorcontrib><collection>CrossRef</collection><jtitle>Transportation research. Part C, Emerging technologies</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Wang, Jiawei</au><au>Sun, Lijun</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Dynamic holding control to avoid bus bunching: A multi-agent deep reinforcement learning framework</atitle><jtitle>Transportation research. Part C, Emerging technologies</jtitle><date>2020-07-01</date><risdate>2020</risdate><volume>116</volume><spage>102661</spage><pages>102661-</pages><artnum>102661</artnum><issn>0968-090X</issn><eissn>1879-2359</eissn><abstract>•A multi-agent deep reinforcement learning framework is proposed for bus holding control.•A reward function is defined to achieve headway self-equalization.•The action of each agent is considered by introducing a joint-action tracker.•A scheme based on proximal policy optimization is designed to train the agents.•The framework outperforms other baselines in simulation studies. Bus bunching has been a long-standing problem that undermines the efficiency and reliability of public transport services. The most popular countermeasure in practice is to introduce static and dynamic holding control. However, most previous holding control strategies mainly consider local information with a pre-specified headway/schedule, while the global coordination of the whole bus fleet and its long-term effect are often overlooked. To efficiently incorporate global coordination and long-term operation in bus holding, in this paper we propose a multi-agent deep reinforcement learning (MDRL) framework to develop dynamic and flexible holding control strategies for a bus route. Specifically, we model each bus as an agent that interacts with not only its leader/follower but also all other vehicles in the fleet. To better explore potential strategies, we develop an effective headway-based reward function in the proposed framework. In the learning framework, we model fleet coordination by using a basic actor-critic scheme along with a joint action tracker to better characterize the complex interactions among agents in policy learning, and we apply proximal policy optimization to improve learning performance. We conduct extensive numerical experiments to evaluate the proposed MDRL framework against multiple baseline models that only rely on local information. Our results demonstrate the superiority of the proposed framework and show the promise of applying MDRL in the coordinative control of public transport vehicle fleets in real-world operations.</abstract><pub>Elsevier Ltd</pub><doi>10.1016/j.trc.2020.102661</doi></addata></record>
fulltext fulltext
identifier ISSN: 0968-090X
ispartof Transportation research. Part C, Emerging technologies, 2020-07, Vol.116, p.102661, Article 102661
issn 0968-090X
1879-2359
language eng
recordid cdi_crossref_primary_10_1016_j_trc_2020_102661
source ScienceDirect Freedom Collection
subjects Bus bunching
Deep reinforcement learning
Dynamic holding
Multi-agent system
Policy proximal optimization
title Dynamic holding control to avoid bus bunching: A multi-agent deep reinforcement learning framework
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-23T16%3A12%3A29IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-elsevier_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Dynamic%20holding%20control%20to%20avoid%20bus%20bunching:%20A%20multi-agent%20deep%20reinforcement%20learning%20framework&rft.jtitle=Transportation%20research.%20Part%20C,%20Emerging%20technologies&rft.au=Wang,%20Jiawei&rft.date=2020-07-01&rft.volume=116&rft.spage=102661&rft.pages=102661-&rft.artnum=102661&rft.issn=0968-090X&rft.eissn=1879-2359&rft_id=info:doi/10.1016/j.trc.2020.102661&rft_dat=%3Celsevier_cross%3ES0968090X20305763%3C/elsevier_cross%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c325t-a9de5e2071c4b7fdbcc565523b0cb8787afc2e0e3acb51a71527ae8b472e6afc3%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true