🤖 AI Summary
This work addresses the joint optimization of long-term utility maximization and strict Quality-of-Service (QoS) constraints in wireless networks. Methodologically, we propose a stochastic policy learning framework based on Generative Diffusion Models (GDMs)—the first application of GDMs to wireless resource allocation—where a supervised policy-learning paradigm imitates expert solutions; Graph Neural Networks (GNNs) encode network topology for cross-topology generalization; and sequential sampling approximates the optimal policy. Evaluated on multi-user interference channel power control, our approach significantly outperforms conventional reinforcement learning and heuristic methods. It achieves near-optimal long-term utility while rigorously satisfying hard QoS constraints. The framework combines theoretical soundness—grounded in diffusion-based policy optimization and graph representation learning—with practical deployability, enabling scalable, topology-agnostic resource management in dynamic wireless environments.
📝 Abstract
This paper proposes a supervised training algorithm for learning stochastic resource allocation policies with generative diffusion models (GDMs). We formulate the allocation problem as the maximization of an ergodic utility function subject to ergodic Quality of Service (QoS) constraints. Given samples from a stochastic expert policy that yields a near-optimal solution to the problem, we train a GDM policy to imitate the expert and generate new samples from the optimal distribution. We achieve near-optimal performance through sequential execution of the generated samples. To enable generalization to a family of network configurations, we parameterize the backward diffusion process with a graph neural network (GNN) architecture. We present numerical results in a case study of power control in multi-user interference networks.