TY - GEN
T1 - Improvement of Multi-agent Continuous Cooperative Patrolling with Learning of Activity Length
AU - Sugiyama, Ayumi
AU - Wu, Lingying
AU - Sugawara, Toshiharu
N1 - Funding Information:
T. Sugawara?This work was partly supported by JSPS KAKENHI (17KT0044).
Publisher Copyright:
© Springer Nature Switzerland AG 2019.
PY - 2019
Y1 - 2019
N2 - We propose a learning method that decides the period of activity according to environmental characteristics and the behavioral strategies in the multi-agent continuous cooperative patrol problem. With recent advances in computer and sensor technologies, agents, which are intelligent control programs running on computers and robots, obtain high autonomy so that they can operate in various fields without pre-defined knowledge. However, cooperation/coordination between agents is sophisticated and complicated to implement. We focus on the activity cycle length (ACL) which is the time length from when an agent starts a patrol to when the agent returns to a charging base in the context of a cooperative patrol where agents, like robots, have batteries with limited capacity. A long ACL will enable an agent to visit distant locations, but the agent will require a long rest time to recharge. The basic idea of our method is that if agents have long-life batteries, they can appropriately shorten the ACL, and thus can visit important locations with a short interval of time by recharging frequently. However, appropriate ACL must depend on many elements such as environmental size, number of agents, workload in an environment, and other agents’ behavior and ACLs. Therefore, we propose a method in which agents autonomously learn the appropriate ACL on the basis of the number of events detected per cycle. We experimentally indicate that our agents are able to learn appropriate ACL depending on established spatial divisional cooperation. We also report the details of the analysis of the experimental results to understand the behaviors of agents with different ACLs.
AB - We propose a learning method that decides the period of activity according to environmental characteristics and the behavioral strategies in the multi-agent continuous cooperative patrol problem. With recent advances in computer and sensor technologies, agents, which are intelligent control programs running on computers and robots, obtain high autonomy so that they can operate in various fields without pre-defined knowledge. However, cooperation/coordination between agents is sophisticated and complicated to implement. We focus on the activity cycle length (ACL) which is the time length from when an agent starts a patrol to when the agent returns to a charging base in the context of a cooperative patrol where agents, like robots, have batteries with limited capacity. A long ACL will enable an agent to visit distant locations, but the agent will require a long rest time to recharge. The basic idea of our method is that if agents have long-life batteries, they can appropriately shorten the ACL, and thus can visit important locations with a short interval of time by recharging frequently. However, appropriate ACL must depend on many elements such as environmental size, number of agents, workload in an environment, and other agents’ behavior and ACLs. Therefore, we propose a method in which agents autonomously learn the appropriate ACL on the basis of the number of events detected per cycle. We experimentally indicate that our agents are able to learn appropriate ACL depending on established spatial divisional cooperation. We also report the details of the analysis of the experimental results to understand the behaviors of agents with different ACLs.
KW - Battery limitation
KW - Continuous cooperative patrol problem
KW - Cycle learning
KW - Division of labor
KW - Multi-agent
UR - http://www.scopus.com/inward/record.url?scp=85077498542&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85077498542&partnerID=8YFLogxK
U2 - 10.1007/978-3-030-37494-5_14
DO - 10.1007/978-3-030-37494-5_14
M3 - Conference contribution
AN - SCOPUS:85077498542
SN - 9783030374938
T3 - Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
SP - 270
EP - 292
BT - Agents and Artificial Intelligence - 11th International Conference, ICAART 2019, Revised Selected Papers
A2 - van den Herik, Jaap
A2 - Rocha, Ana Paula
A2 - Steels, Luc
PB - Springer
T2 - 11th International Conference on Agents and Artificial Intelligence, ICAART 2019
Y2 - 19 February 2019 through 21 February 2019
ER -