🤖 AI Summary
While causal reasoning has been extensively studied in single-agent reinforcement learning (RL), its applicability and challenges in multi-agent RL (MARL) remain largely unexplored, with no prior work establishing causal modeling frameworks for collaborative MARL. Method: This paper introduces the first causal-enhanced framework for cooperative MARL, integrating causal discovery (via the PC algorithm), structural causal models (SCMs), and mainstream MARL algorithms (MAPPO and QMix) to perform causal state augmentation in cooperative tasks. Contribution/Results: Experiments demonstrate that causal enhancement significantly improves policy robustness and sample efficiency—especially in high-cooperation settings—and uncover novel challenges, including causal identification ambiguity in multi-agent systems. The paper further proposes the first research roadmap for causal modeling in MARL. Collectively, this work establishes both a methodological foundation and empirical evidence for causal MARL.
📝 Abstract
Causal reasoning is increasingly used in Reinforcement Learning (RL) to improve the learning process in several dimensions: efficacy of learned policies, efficiency of convergence, generalisation capabilities, safety and interpretability of behaviour. However, applications of causal reasoning to Multi-Agent RL (MARL) are still mostly unexplored. In this paper, we take the first step in investigating the opportunities and challenges of applying causal reasoning in MARL. We measure the impact of a simple form of causal augmentation in state-of-the-art MARL scenarios increasingly requiring cooperation, and with state-of-the-art MARL algorithms exploiting various degrees of collaboration between agents. Then, we discuss the positive as well as negative results achieved, giving us the chance to outline the areas where further research may help to successfully transfer causal RL to the multi-agent setting.