Random Latent Exploration for Deep Reinforcement Learning

๐Ÿ“… 2024-07-18
๐Ÿ›๏ธ International Conference on Machine Learning
๐Ÿ“ˆ Citations: 1
โœจ Influential: 0
๐Ÿ“„ PDF
๐Ÿค– AI Summary
This work addresses insufficient exploration in deep reinforcement learning under sparse-reward settings. We propose a reward-free exploration method guided by stochastic latent-space goals. Our approach uniquely integrates random latent goal sampling with reward-free policy shapingโ€”achieving intrinsic-motivation-like deep exploration without introducing any auxiliary intrinsic reward module. By jointly modeling the latent space and learning goal-conditioned policies, we construct an end-to-end, plug-and-play exploration component. Evaluated on both discrete-action Atari and continuous-control Isaac Gym benchmarks, our method consistently outperforms standard noise-injection baselines (e.g., parameter or action space noise) and reward-shaping approaches in average performance. Crucially, it achieves this while maintaining architectural simplicity and strong generalization across diverse environments and tasks.

Technology Category

Application Category

๐Ÿ“ Abstract
We introduce Random Latent Exploration (RLE), a simple yet effective exploration strategy in reinforcement learning (RL). On average, RLE outperforms noise-based methods, which perturb the agent's actions, and bonus-based exploration, which rewards the agent for attempting novel behaviors. The core idea of RLE is to encourage the agent to explore different parts of the environment by pursuing randomly sampled goals in a latent space. RLE is as simple as noise-based methods, as it avoids complex bonus calculations but retains the deep exploration benefits of bonus-based methods. Our experiments show that RLE improves performance on average in both discrete (e.g., Atari) and continuous control tasks (e.g., Isaac Gym), enhancing exploration while remaining a simple and general plug-in for existing RL algorithms.
Problem

Research questions and friction points this paper is trying to address.

Enhance exploration in reinforcement learning
Simplify exploration strategy complexity
Improve performance in control tasks
Innovation

Methods, ideas, or system contributions that make the work stand out.

Random Latent Exploration strategy
Simplifies exploration in RL
Enhances performance in control tasks
๐Ÿ”Ž Similar Papers
No similar papers found.