Geometry of Neural Reinforcement Learning in Continuous State and Action Spaces

πŸ“… 2025-07-28
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
This paper addresses the weak theoretical foundations of reinforcement learning (RL) in continuous state-action spaces. To this end, it introduces a geometric analytical framework that constructs a manifold of state reachability induced by stochastic policies. This framework establishes, for the first time, a rigorous theoretical connection between the geometric structure of the state space and the dimensionality of the action space: it proves that the intrinsic dimension of the reachable manifold scales withβ€”and is upper-bounded byβ€”the action-space dimension. Methodologically, the approach integrates a local manifold-learning layer into a two-layer neural policy network, embedded within an actor-critic architecture and trained via semi-gradient optimization to learn sparse, low-dimensional state representations. Experiments on MuJoCo benchmarks and synthetic environments demonstrate substantial improvements in sample efficiency and policy performance for high-degree-of-freedom control tasks.

Technology Category

Application Category

πŸ“ Abstract
Advances in reinforcement learning (RL) have led to its successful application in complex tasks with continuous state and action spaces. Despite these advances in practice, most theoretical work pertains to finite state and action spaces. We propose building a theoretical understanding of continuous state and action spaces by employing a geometric lens to understand the locally attained set of states. The set of all parametrised policies learnt through a semi-gradient based approach induces a set of attainable states in RL. We show that the training dynamics of a two-layer neural policy induce a low dimensional manifold of attainable states embedded in the high-dimensional nominal state space trained using an actor-critic algorithm. We prove that, under certain conditions, the dimensionality of this manifold is of the order of the dimensionality of the action space. This is the first result of its kind, linking the geometry of the state space to the dimensionality of the action space. We empirically corroborate this upper bound for four MuJoCo environments and also demonstrate the results in a toy environment with varying dimensionality. We also show the applicability of this theoretical result by introducing a local manifold learning layer to the policy and value function networks to improve the performance in control environments with very high degrees of freedom by changing one layer of the neural network to learn sparse representations.
Problem

Research questions and friction points this paper is trying to address.

Theoretical understanding of continuous state and action spaces in RL
Geometry of attainable states in high-dimensional nominal state space
Linking state space geometry to action space dimensionality
Innovation

Methods, ideas, or system contributions that make the work stand out.

Geometric lens for continuous RL analysis
Low-dimensional manifold in state space
Local manifold learning layer enhancement
πŸ”Ž Similar Papers
No similar papers found.
Saket Tiwari
Saket Tiwari
PhD Candidate, Brown University
Reinforcement LearningMachine LearningDeep Learning
O
Omer Gottesman
Amazon Web Services
G
George Konidaris
Department of Computer Science, Brown University