TY - JOUR
T1 - Coordinated control method for ridesharing service area using deep reinforcement learning
AU - Yoshida, Naoki
AU - Noda, Itsuki
AU - Sugawara, Toshiharu
N1 - Publisher Copyright:
© 2021, Japanese Society for Artificial Intelligence. All rights reserved.
PY - 2021
Y1 - 2021
N2 - We propose a coordinated control method of agents, which are self-driving ridesharing vehicles, by using multi-agent deep reinforcement learning (MADRL) so that they individually determine where they should wait for passengers to provide better services as well as to increase their profits in rideshare services. With the increasing demand for ridesharing services, many drivers and passengers have started to participate. However, many drivers spend most of their operating time with empty vehicles, which is not only inefficient but also causes problems such as wasted energy, increased traffic congestion in urban areas, and shortages of ridesharing vehicles in less demand areas. To address this issue, distributed service area adaptation method for ride sharing (dSAAMS), in which agents learn where they should wait using MADRL, was already proposed, but we found that it does not work well under certain environments. Therefore, we propose dSAAMS* with modified input and improved reward scheme for agents to generate coordinated behaviors to adapt to various environments. Then, we evaluated the performance and charac-teristics of the proposed method by using a simulation environment with varying passenger generation patterns and real data in Manhattan. Our results indicate that the dSAAMS* provides better quality service than the conventional methods and performs better in dynamically changing environments.
AB - We propose a coordinated control method of agents, which are self-driving ridesharing vehicles, by using multi-agent deep reinforcement learning (MADRL) so that they individually determine where they should wait for passengers to provide better services as well as to increase their profits in rideshare services. With the increasing demand for ridesharing services, many drivers and passengers have started to participate. However, many drivers spend most of their operating time with empty vehicles, which is not only inefficient but also causes problems such as wasted energy, increased traffic congestion in urban areas, and shortages of ridesharing vehicles in less demand areas. To address this issue, distributed service area adaptation method for ride sharing (dSAAMS), in which agents learn where they should wait using MADRL, was already proposed, but we found that it does not work well under certain environments. Therefore, we propose dSAAMS* with modified input and improved reward scheme for agents to generate coordinated behaviors to adapt to various environments. Then, we evaluated the performance and charac-teristics of the proposed method by using a simulation environment with varying passenger generation patterns and real data in Manhattan. Our results indicate that the dSAAMS* provides better quality service than the conventional methods and performs better in dynamically changing environments.
KW - Deep reinforcement learning
KW - Multi-agent learning
KW - Ride sharing
KW - Transportation and logistics
UR - http://www.scopus.com/inward/record.url?scp=85114273383&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85114273383&partnerID=8YFLogxK
U2 - 10.1527/tjsai.36-5_AG21-D
DO - 10.1527/tjsai.36-5_AG21-D
M3 - Article
AN - SCOPUS:85114273383
SN - 1346-0714
VL - 36
JO - Transactions of the Japanese Society for Artificial Intelligence
JF - Transactions of the Japanese Society for Artificial Intelligence
IS - 5
M1 - AG21-D_1
ER -