🤖 AI Summary
This study addresses the joint real-time decoding of auditory attention direction (spatial location) and timbre attributes from EEG signals under realistic noisy conditions. To eliminate pre-trial information leakage, we introduce a cue-masked auditory attention paradigm. We further propose AADNet—a novel end-to-end deep network integrating CNN and LSTM modules—capable of simultaneously modeling spatiotemporal EEG features within a mere 0.5-second window, without requiring access to original audio stimuli or prior knowledge. To our knowledge, this is the first work achieving high-accuracy parallel decoding of both attention direction and timbre, attaining mean classification accuracies of 93.46% and 91.09%, respectively—significantly outperforming five state-of-the-art baseline methods. The framework enables real-time, multi-attribute attention recognition for closed-loop neurostimulation-based hearing aids, thereby advancing the practical deployment of brain–computer interfaces in complex acoustic environments.
📝 Abstract
Auditory attention decoding from electroencephalogram (EEG) could infer to which source the user is attending in noisy environments. Decoding algorithms and experimental paradigm designs are crucial for the development of technology in practical applications. To simulate real-world scenarios, this study proposed a cue-masked auditory attention paradigm to avoid information leakage before the experiment. To obtain high decoding accuracy with low latency, an end-to-end deep learning model, AADNet, was proposed to exploit the spatiotemporal information from the short time window of EEG signals. The results showed that with a 0.5-second EEG window, AADNet achieved an average accuracy of 93.46% and 91.09% in decoding auditory orientation attention (OA) and timbre attention (TA), respectively. It significantly outperformed five previous methods and did not need the knowledge of the original audio source. This work demonstrated that it was possible to detect the orientation and timbre of auditory attention from EEG signals fast and accurately. The results are promising for the real-time multi-property auditory attention decoding, facilitating the application of the neuro-steered hearing aids and other assistive listening devices.