π€ AI Summary
The critical frequency characteristics underlying native-language speech and music envelope-following responses (EFRs) in auditory cognition remain poorly understood.
Method: Using multichannel EEG recordings, we applied time-averaged spectral response and cross-spectral density analyses within a linear time-varying system framework to estimate neural transfer functions, thereby quantifying the frequency-domain mapping between acoustic envelopes and cortical responses.
Contribution/Results: We identified three robust, scalp-wide neural coherence peaks at 10β13 Hz (alpha), 27β29 Hz (low gamma), and 62β64 Hz (high gamma)βthe first such demonstration across stimulus domains. These bands correspond respectively to sustained attention, acoustic feature binding, and working memory processing. Our findings demonstrate that the auditory pathway achieves efficient neural synchronization to both speech and music via oscillatory activity in these discrete frequency bands, providing novel evidence for rhythmic coding mechanisms in auditory perception.
π Abstract
Although native speech and music envelope following responses (EFRs) play a crucial role in auditory processing and cognition, their frequency profile, such as the dominating frequency and spectral coherence, is largely unknown. We have assumed that the auditory pathway - which transmits envelope components of speech and music to the scalp through time-varying neurophysiological processes - is a linear time-varying system, with the envelope and the multi-channel EEG responses as excitation and response, respectively. This paper investigates the transfer function of this system through two analytical techniques - time-averaged spectral responses and cross-spectral density - in the frequency domain at four different positions of the human scalp. Our findings suggest that alpha (8-11 Hz), lower gamma (53-56 Hz), and higher gamma (78-81 Hz) bands are the peak responses of the system. These frequently appearing dominant frequency responses may be the key components of familiar speech perception, maintaining attention, binding acoustic features, and memory processing. The cross-spectral density, which reflects the spatial neural coherence of the human brain, shows that 10-13 Hz, 27-29 Hz, and 62-64 Hz are common for all channel pairs. As neural coherences are frequently observed in these frequencies among native participants, we suggest that these distributed neural processes are also dominant in native speech and music perception.