TY - GEN
T1 - Learning to coordinate with deep reinforcement learning in doubles pong game
AU - Diallo, Elhadji Amadou Oury
AU - Sugiyama, Ayumi
AU - Sugawara, Toshiharu
N1 - Publisher Copyright:
© 2017 IEEE.
PY - 2017
Y1 - 2017
N2 - This paper discusses the emergence of cooperative and coordinated behaviors between joint and concurrent learning agents using deep Q-learning. Multi-agent systems (MAS) arise in a variety of domains. The collective effort is one of the main building blocks of many fundamental systems that exist in the world, and thus, sequential decision making under uncertainty for collaborative work is one of the important and challenging issues for intelligent cooperative multiple agents. However, the decisions for cooperation are highly sophisticated and complicated because agents may have a certain shared goal or individual goals to achieve and their behavior is inevitably influenced by each other. Therefore, we attempt to explore whether agents using deep Q-networks (DQN) can learn cooperative behavior. We use doubles pong game as an example and we investigate how they learn to divide their works through iterated game executions. In our approach, agents jointly learn to divide their area of responsibility and each agent uses its own DQN to modify its behavior. We also investigate how learned behavior changes according to environmental characteristics including reward schemes and learning techniques. Our experiments indicate that effective cooperative behaviors with balanced division of workload emerge. These results help us to better understand how agents behave and interact with each other in complex environments and how they coherently choose their individual actions such that the resulting joint actions are optimal.
AB - This paper discusses the emergence of cooperative and coordinated behaviors between joint and concurrent learning agents using deep Q-learning. Multi-agent systems (MAS) arise in a variety of domains. The collective effort is one of the main building blocks of many fundamental systems that exist in the world, and thus, sequential decision making under uncertainty for collaborative work is one of the important and challenging issues for intelligent cooperative multiple agents. However, the decisions for cooperation are highly sophisticated and complicated because agents may have a certain shared goal or individual goals to achieve and their behavior is inevitably influenced by each other. Therefore, we attempt to explore whether agents using deep Q-networks (DQN) can learn cooperative behavior. We use doubles pong game as an example and we investigate how they learn to divide their works through iterated game executions. In our approach, agents jointly learn to divide their area of responsibility and each agent uses its own DQN to modify its behavior. We also investigate how learned behavior changes according to environmental characteristics including reward schemes and learning techniques. Our experiments indicate that effective cooperative behaviors with balanced division of workload emerge. These results help us to better understand how agents behave and interact with each other in complex environments and how they coherently choose their individual actions such that the resulting joint actions are optimal.
KW - Artificial Intelligence
KW - Cooperative Systems
KW - Deep Reinforcement Learning
KW - Multi agent systems
UR - http://www.scopus.com/inward/record.url?scp=85048507667&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85048507667&partnerID=8YFLogxK
U2 - 10.1109/ICMLA.2017.0-184
DO - 10.1109/ICMLA.2017.0-184
M3 - Conference contribution
AN - SCOPUS:85048507667
T3 - Proceedings - 16th IEEE International Conference on Machine Learning and Applications, ICMLA 2017
SP - 14
EP - 19
BT - Proceedings - 16th IEEE International Conference on Machine Learning and Applications, ICMLA 2017
A2 - Chen, Xuewen
A2 - Luo, Bo
A2 - Luo, Feng
A2 - Palade, Vasile
A2 - Wani, M. Arif
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 16th IEEE International Conference on Machine Learning and Applications, ICMLA 2017
Y2 - 18 December 2017 through 21 December 2017
ER -