Adaptive Parameter Sharing for Multi-Agent Reinforcement Learning

📅 2023-12-14
🏛️ IEEE International Conference on Acoustics, Speech, and Signal Processing
📈 Citations: 2
Influential: 1
📄 PDF
🤖 AI Summary
To address policy homogenization induced by parameter sharing in multi-agent reinforcement learning—particularly its inability to accommodate heterogeneous agent identities and task requirements—this paper proposes a zero-overhead, identity-driven adaptive subnet partitioning mechanism. Inspired by neural functional parcellation, the method employs a learnable identity encoder to generate agent-specific binary masks, which dynamically route inputs to localized subnetworks within a shared backbone, thereby enabling differentiated policy representations. Crucially, it introduces no additional parameters and is fully compatible with standard on-policy algorithms such as PPO and A2C. Extensive experiments on StarCraft II, the Multi-Agent Particle Environment (MPE), and custom heterogeneous benchmarks demonstrate an average 12.7% improvement in win rate and a 3.2× increase in inter-agent policy diversity, significantly outperforming both conventional parameter sharing and Hypernetwork-based baselines.
📝 Abstract
Parameter sharing, as an important technique in multi-agent systems, can effectively solve the scalability issue in large-scale agent problems. However, the effectiveness of parameter sharing largely depends on the environment setting. When agents have different identities or tasks, naive parameter sharing makes it difficult to generate sufficiently differentiated strategies for agents. Inspired by research pertaining to the brain in biology, we propose a novel parameter sharing method. It maps each type of agent to different regions within a shared network based on their identity, resulting in distinct subnetworks. Therefore, our method can increase the diversity of strategies among different agents without introducing additional training parameters. Through experiments conducted in multiple environments, our method has shown better performance than other parameter sharing methods.
Problem

Research questions and friction points this paper is trying to address.

Addresses scalability in multi-agent reinforcement learning
Enhances strategy diversity among agents with different identities
Improves performance without additional training parameters
Innovation

Methods, ideas, or system contributions that make the work stand out.

Adaptive parameter sharing for multi-agent systems
Maps agent types to distinct network regions
Increases strategy diversity without extra parameters
🔎 Similar Papers
No similar papers found.
Dapeng Li
Dapeng Li
Institute of Automation, Chinese Academy of Sciences
MARLLLM
N
Na Lou
Institute of Automation, Chinese Academy of Sciences; School of Artificial Intelligence, University of Chinese Academy of Sciences
B
Bin Zhang
Institute of Automation, Chinese Academy of Sciences; School of Artificial Intelligence, University of Chinese Academy of Sciences
Z
Zhiwei Xu
Institute of Automation, Chinese Academy of Sciences; School of Artificial Intelligence, University of Chinese Academy of Sciences
Guoliang Fan
Guoliang Fan
Professor of Electrical Engineering at Oklahoma State University
image processingcomputer visionmachine learningmultimedia