TY - JOUR
T1 - Analysis of coordinated behavior structures with multi-agent deep reinforcement learning
AU - Miyashita, Yuki
AU - Sugawara, Toshiharu
N1 - Funding Information:
This work was partly supported by JSPS KAKENHI Grant Number 17KT0044.
Publisher Copyright:
© 2020, The Author(s).
PY - 2021/2
Y1 - 2021/2
N2 - Cooperation and coordination are major issues in studies on multi-agent systems because the entire performance of such systems is greatly affected by these activities. The issues are challenging however, because appropriate coordinated behaviors depend on not only environmental characteristics but also other agents’ strategies. On the other hand, advances in multi-agent deep reinforcement learning (MADRL) have recently attracted attention, because MADRL can considerably improve the entire performance of multi-agent systems in certain domains. The characteristics of learned coordination structures and agent’s resulting behaviors, however, have not been clarified sufficiently. Therefore, we focus here on MADRL in which agents have their own deep Q-networks (DQNs), and we analyze their coordinated behaviors and structures for the pickup and floor laying problem, which is an abstraction of our target application. In particular, we analyze the behaviors around scarce resources and long narrow passages in which conflicts such as collisions are likely to occur. We then indicated that different types of inputs to the networks exhibit similar performance but generate various coordination structures with associated behaviors, such as division of labor and a shared social norm, with no direct communication.
AB - Cooperation and coordination are major issues in studies on multi-agent systems because the entire performance of such systems is greatly affected by these activities. The issues are challenging however, because appropriate coordinated behaviors depend on not only environmental characteristics but also other agents’ strategies. On the other hand, advances in multi-agent deep reinforcement learning (MADRL) have recently attracted attention, because MADRL can considerably improve the entire performance of multi-agent systems in certain domains. The characteristics of learned coordination structures and agent’s resulting behaviors, however, have not been clarified sufficiently. Therefore, we focus here on MADRL in which agents have their own deep Q-networks (DQNs), and we analyze their coordinated behaviors and structures for the pickup and floor laying problem, which is an abstraction of our target application. In particular, we analyze the behaviors around scarce resources and long narrow passages in which conflicts such as collisions are likely to occur. We then indicated that different types of inputs to the networks exhibit similar performance but generate various coordination structures with associated behaviors, such as division of labor and a shared social norm, with no direct communication.
KW - Cooperation
KW - Coordination
KW - Divisional cooperation
KW - Multi-agent deep reinforcement learning
KW - Norm
UR - http://www.scopus.com/inward/record.url?scp=85091071781&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85091071781&partnerID=8YFLogxK
U2 - 10.1007/s10489-020-01832-y
DO - 10.1007/s10489-020-01832-y
M3 - Article
AN - SCOPUS:85091071781
SN - 0924-669X
VL - 51
SP - 1069
EP - 1085
JO - Applied Intelligence
JF - Applied Intelligence
IS - 2
ER -