π€ AI Summary
Existing deep reinforcement learning (DRL) approaches for socially aware navigation of mobile robots in dynamic pedestrian environments suffer from limited action flexibility due to restrictive Gaussian distribution assumptions. To address this, we propose the first diffusion-model-based framework for social navigation RLβreplacing fixed-distribution priors with a flexible, controllable generative process for continuous action synthesis and real-time adaptation. Our method enables post-training action smoothing optimization and zero-shot transfer to static obstacle scenarios, significantly enhancing generalization and robustness. Experiments demonstrate a 12.7% improvement in navigation success rate over Gaussian-policy baselines in dense dynamic environments, while maintaining robust performance on static obstacles without fine-tuning.
π Abstract
Mobile robot navigation in dynamic environments with pedestrian traffic is a key challenge in the development of autonomous mobile service robots. Recently, deep reinforcement learning-based methods have been actively studied and have outperformed traditional rule-based approaches owing to their optimization capabilities. Among these, methods that assume a continuous action space typically rely on a Gaussian distribution assumption, which limits the flexibility of generated actions. Meanwhile, the application of diffusion models to reinforcement learning has advanced, allowing for more flexible action distributions compared with Gaussian distribution-based approaches. In this study, we applied a diffusion-based reinforcement learning approach to social navigation and validated its effectiveness. Furthermore, by leveraging the characteristics of diffusion models, we propose an extension that enables post-training action smoothing and adaptation to static obstacle scenarios not considered during the training steps.