🤖 AI Summary
Deep learning–based EEG decoding often suffers from latent overfitting, severely impairing cross-scenario generalization and hindering real-world BCI deployment. To address this, we propose the weakly supervised Disentangled Decoding Decomposition (D3) framework. D3 leverages positional prediction within trial sequences to drive latent-space disentanglement; introduces a novel independent subnetwork architecture coupled with a nonlinear ICA–inspired disentanglement mechanism to isolate task-agnostic neural components; and establishes a cross-dataset component activation contrast paradigm that explicitly models linear separability in the latent space. Experiments demonstrate that D3 stably isolates physiologically meaningful components in motor imagery data while suppressing task-irrelevant artifacts. In sleep staging, D3 achieves significant few-shot accuracy gains—requiring only minimal labeled data—and generalizes robustly to real-world scenarios.
📝 Abstract
Deep learning for decoding EEG signals has gained traction, with many claims to state-of-the-art accuracy. However, despite the convincing benchmark performance, successful translation to real applications is limited. The frequent disconnect between performance on controlled BCI benchmarks and its lack of generalisation to practical settings indicates hidden overfitting problems. We introduce Disentangled Decoding Decomposition (D3), a weakly supervised method for training deep learning models across EEG datasets. By predicting the place in the respective trial sequence from which the input window was sampled, EEG-D3 separates latent components of brain activity, akin to non-linear ICA. We utilise a novel model architecture with fully independent sub-networks for strict interpretability. We outline a feature interpretation paradigm to contrast the component activation profiles on different datasets and inspect the associated temporal and spatial filters. The proposed method reliably separates latent components of brain activity on motor imagery data. Training downstream classifiers on an appropriate subset of these components prevents hidden overfitting caused by task-correlated artefacts, which severely affects end-to-end classifiers. We further exploit the linearly separable latent space for effective few-shot learning on sleep stage classification. The ability to distinguish genuine components of brain activity from spurious features results in models that avoid the hidden overfitting problem and generalise well to real-world applications, while requiring only minimal labelled data. With interest to the neuroscience community, the proposed method gives researchers a tool to separate individual brain processes and potentially even uncover heretofore unknown dynamics.