Where is the cocktail party? Decoding locations of attended and unattended moving sound sources using EEG
- PMID: 31629828
- DOI: 10.1016/j.neuroimage.2019.116283
Where is the cocktail party? Decoding locations of attended and unattended moving sound sources using EEG
Abstract
Recently, we showed that in a simple acoustic scene with one sound source, auditory cortex tracks the time-varying location of a continuously moving sound. Specifically, we found that both the delta phase and alpha power of the electroencephalogram (EEG) can be used to reconstruct the sound source azimuth. However, in natural settings, we are often presented with a mixture of multiple competing sounds and so we must focus our attention on the relevant source in order to segregate it from the competing sources e.g. 'cocktail party effect'. While many studies have examined this phenomenon in the context of sound envelope tracking by the cortex, it is unclear how we process and utilize spatial information in complex acoustic scenes with multiple sound sources. To test this, we created an experiment where subjects listened to two concurrent sound stimuli that were moving within the horizontal plane over headphones while we recorded their EEG. Participants were tasked with paying attention to one of the two presented stimuli. The data were analyzed by deriving linear mappings, temporal response functions (TRF), between EEG data and attended as well unattended sound source trajectories. Next, we used these TRFs to reconstruct both trajectories from previously unseen EEG data. In a first experiment we used noise stimuli and included the task involved spatially localizing embedded targets. Then, in a second experiment, we employed speech stimuli and a non-spatial speech comprehension task. Results showed the trajectory of an attended sound source can be reliably reconstructed from both delta phase and alpha power of EEG even in the presence of distracting stimuli. Moreover, the reconstruction was robust to task and stimulus type. The cortical representation of the unattended source position was below detection level for the noise stimuli, but we observed weak tracking of the unattended source location for the speech stimuli by the delta phase of EEG. In addition, we demonstrated that the trajectory reconstruction method can in principle be used to decode selective attention on a single-trial basis, however, its performance was inferior to envelope-based decoders. These results suggest a possible dissociation of delta phase and alpha power of EEG in the context of sound trajectory tracking. Moreover, the demonstrated ability to localize and determine the attended speaker in complex acoustic environments is particularly relevant for cognitively controlled hearing devices.
Keywords: Attention; Cocktail party; Electroencephalography; Localization; Motion; Sound.
Copyright © 2019 Elsevier Inc. All rights reserved.
Similar articles
-
Neural tracking of auditory motion is reflected by delta phase and alpha power of EEG.Neuroimage. 2018 Nov 1;181:683-691. doi: 10.1016/j.neuroimage.2018.07.054. Epub 2018 Jul 24. Neuroimage. 2018. PMID: 30053517
-
Noise-robust cortical tracking of attended speech in real-world acoustic scenes.Neuroimage. 2017 Aug 1;156:435-444. doi: 10.1016/j.neuroimage.2017.04.026. Epub 2017 Apr 13. Neuroimage. 2017. PMID: 28412441
-
EEG decoding of the target speaker in a cocktail party scenario: considerations regarding dynamic switching of talker location.J Neural Eng. 2019 Jun;16(3):036017. doi: 10.1088/1741-2552/ab0cf1. Epub 2019 Mar 5. J Neural Eng. 2019. PMID: 30836345
-
Evidence against attentional state modulating scalp-recorded auditory brainstem steady-state responses.Brain Res. 2015 Nov 11;1626:146-64. doi: 10.1016/j.brainres.2015.06.038. Epub 2015 Jul 14. Brain Res. 2015. PMID: 26187756 Free PMC article. Review.
-
Hearing in Complex Environments: Auditory Gain Control, Attention, and Hearing Loss.Front Neurosci. 2022 Feb 10;16:799787. doi: 10.3389/fnins.2022.799787. eCollection 2022. Front Neurosci. 2022. PMID: 35221899 Free PMC article. Review.
Cited by
-
Neural representation of linguistic feature hierarchy reflects second-language proficiency.Neuroimage. 2021 Feb 15;227:117586. doi: 10.1016/j.neuroimage.2020.117586. Epub 2020 Dec 24. Neuroimage. 2021. PMID: 33346131 Free PMC article.
-
Convolutional neural networks can identify brain interactions involved in decoding spatial auditory attention.PLoS Comput Biol. 2024 Aug 8;20(8):e1012376. doi: 10.1371/journal.pcbi.1012376. eCollection 2024 Aug. PLoS Comput Biol. 2024. PMID: 39116183 Free PMC article.
-
Auditory Attention Detection via Cross-Modal Attention.Front Neurosci. 2021 Jul 21;15:652058. doi: 10.3389/fnins.2021.652058. eCollection 2021. Front Neurosci. 2021. PMID: 34366770 Free PMC article.
-
A GRU-CNN model for auditory attention detection using microstate and recurrence quantification analysis.Sci Rep. 2024 Apr 17;14(1):8861. doi: 10.1038/s41598-024-58886-y. Sci Rep. 2024. PMID: 38632246 Free PMC article.
-
A comprehensive multimodal dataset for contactless lip reading and acoustic analysis.Sci Data. 2023 Dec 13;10(1):895. doi: 10.1038/s41597-023-02793-w. Sci Data. 2023. PMID: 38092796 Free PMC article.
Publication types
MeSH terms
LinkOut - more resources
Full Text Sources
Other Literature Sources
