Interpretable Reinforcement Learning for Load Balancing using Kolmogorov-Arnold Networks

📅 2025-05-20
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Reinforcement learning (RL)-based network load balancing suffers from poor interpretability (“black-box” policies) and intractable controller equations, hindering verification and deployment. Method: This paper proposes the first integration of Kolmogorov–Arnold Networks (KANs) into the Proximal Policy Optimization (PPO) framework, designing a structured, interpretable Actor network (a single-layer KAN) alongside a standard MLP-based Critic. A multi-objective reward function jointly optimizes throughput utility, packet loss rate, and end-to-end latency. Contribution/Results: The approach significantly improves throughput, reduces packet loss, and lowers latency across diverse network settings. Crucially, the KAN-based Actor enables direct analytical extraction of compact, physically meaningful, explicit load-balancing equations from the trained policy—overcoming the fundamental limitations of uninterpretable RL controllers and non-extractable policies. This establishes a novel paradigm for verifiable, deployable intelligent network control.

Technology Category

Application Category

📝 Abstract
Reinforcement learning (RL) has been increasingly applied to network control problems, such as load balancing. However, existing RL approaches often suffer from lack of interpretability and difficulty in extracting controller equations. In this paper, we propose the use of Kolmogorov-Arnold Networks (KAN) for interpretable RL in network control. We employ a PPO agent with a 1-layer actor KAN model and an MLP Critic network to learn load balancing policies that maximise throughput utility, minimize loss as well as delay. Our approach allows us to extract controller equations from the learned neural networks, providing insights into the decision-making process. We evaluate our approach using different reward functions demonstrating its effectiveness in improving network performance while providing interpretable policies.
Problem

Research questions and friction points this paper is trying to address.

Enhance interpretability in reinforcement learning for load balancing
Extract controller equations from learned neural networks
Improve network performance with interpretable policies
Innovation

Methods, ideas, or system contributions that make the work stand out.

Uses Kolmogorov-Arnold Networks for interpretable RL
Combines PPO agent with KAN actor and MLP critic
Extracts controller equations from learned neural networks
🔎 Similar Papers
No similar papers found.
K
Kamal Singh
Universit ´e Jean Monnet Saint- ´Etienne, CNRS, Inst. d’Optique Graduate School, Lab. Hubert Curien, F-42023 Saint- ´Etienne, France
S
Sami Marouani
Universit ´e Jean Monnet Saint- ´Etienne, CNRS, Inst. d’Optique Graduate School, Lab. Hubert Curien, F-42023 Saint- ´Etienne, France
A
Ahmad Al Sheikh
QoS Design, Toulouse, France
Pham Tran Anh Quang
Pham Tran Anh Quang
Huawei Technologies
NetworkingroutingNFV/SDNQoS/QoE
Amaury Habrard
Amaury Habrard
Professor of Computer Science, University Jean Monnet of Saint-Etienne (France)
machine learning