🤖 AI Summary
How does the human brain spontaneously construct fine-grained, six-dimensional egocentric spatial representations (3D position + 3D orientation) during passive observation of first-person videos?
Method: We employed high-temporal-resolution EEG recordings coupled with deep learning–based decoding models to reconstruct continuous spatial pose trajectories from neural activity—without requiring explicit behavioral tasks. Gradient-based reverse mapping was used to characterize the distributed neural encoding mechanisms.
Contribution/Results: We achieved the first real-time, millisecond-scale decoding of fine-grained spatial pose during passive visual perception. Decoding accuracy significantly improved at 100-Hz stimulus frame rates and strongly correlated with subjective spatial judgments. Crucially, we demonstrate that consistent, behaviorally relevant spatial representations emerge endogenously—even in the absence of active engagement—challenging the classical active/passive dichotomy in spatial cognition. This work establishes a noninvasive BCI paradigm capable of resolving high-order endogenous spatial cognition, offering a novel framework for investigating the neurodynamics of embodied spatial perception.
📝 Abstract
Humans possess a remarkable capacity for spatial cognition, allowing for self-localization even in novel or unfamiliar environments. While hippocampal neurons encoding position and orientation are well documented, the large-scale neural dynamics supporting spatial representation, particularly during naturalistic, passive experience, remain poorly understood. Here, we demonstrate for the first time that non-invasive brain-computer interfaces (BCIs) based on electroencephalography (EEG) can decode spontaneous, fine-grained egocentric 6D pose, comprising three-dimensional position and orientation, during passive viewing of egocentric video. Despite EEG's limited spatial resolution and high signal noise, we find that spatially coherent visual input (i.e., continuous and structured motion) reliably evokes decodable spatial representations, aligning with participants' subjective sense of spatial engagement. Decoding performance further improves when visual input is presented at a frame rate of 100 ms per image, suggesting alignment with intrinsic neural temporal dynamics. Using gradient-based backpropagation through a neural decoding model, we identify distinct EEG channels contributing to position -- and orientation specific -- components, revealing a distributed yet complementary neural encoding scheme. These findings indicate that the brain's spatial systems operate spontaneously and continuously, even under passive conditions, challenging traditional distinctions between active and passive spatial cognition. Our results offer a non-invasive window into the automatic construction of egocentric spatial maps and advance our understanding of how the human mind transforms everyday sensory experience into structured internal representations.