Applying Opponent and Environment Modelling in Decentralised Multi-Agent Reinforcement Learning

17 Pages Posted: 18 Sep 2024

See all articles by Alexander Chernyavskiy

Alexander Chernyavskiy

affiliation not provided to SSRN

Aleksandr Panov

affiliation not provided to SSRN

Aleksey Skrynnik

affiliation not provided to SSRN

Abstract

Multi-agent reinforcement learning (MARL) has recently gained popularity and achieved much success in different kind of games such as zero-sum, cooperative or general-sum games. Nevertheless, the vast majority of modern algorithms assume information sharing during training and, hence, could not be utilised in decentralised applications as well as leverage high-dimensional scenarios and be applied to applications with general or sophisticated reward structure. Thus, due to collecting expenses and sparsity of data in real-world applications it becomes necessary to use world models to model the environment dynamics using latent variables --- i.e. use world model to generate synthetic data for training of MARL algorithms. Therefore, focusing on the paradigm of decentralised training and decentralised execution, we propose an extension to the model-based reinforcement learning approaches leveraging fully decentralised training with planning conditioned on neighbouring co-players' latent representations. Our approach is inspired by the idea of opponent modelling. The method makes the agent learn in joint latent space without need to interact with the environment. We suggest the approach as proof of concept that decentralised model-based algorithms are able to emerge collective behaviour with limited communication during planning, and demonstrate its necessity on iterated matrix games and modified versions of StarCraft Multi-Agent Challenge (SMAC).

Keywords: Social and Multi-agent Reinforcement Learning, Reinforcement Learning in Cognitive Science, Multi-agent Systems, Model-based Reinforcement Learning, Opponent Modelling

Suggested Citation

Chernyavskiy, Alexander and Panov, Aleksandr and Skrynnik, Aleksey, Applying Opponent and Environment Modelling in Decentralised Multi-Agent Reinforcement Learning. Available at SSRN: https://ssrn.com/abstract=4959804 or http://dx.doi.org/10.2139/ssrn.4959804

Alexander Chernyavskiy (Contact Author)

affiliation not provided to SSRN ( email )

No Address Available

Aleksandr Panov

affiliation not provided to SSRN ( email )

No Address Available

Aleksey Skrynnik

affiliation not provided to SSRN ( email )

No Address Available

Do you have a job opening that you would like to promote on SSRN?

Paper statistics

Downloads
35
Abstract Views
129
PlumX Metrics