Comparison of Two-Talker Attention Decoding from EEG with Nonlinear Neural Networks and Linear Methods
- PMID: 31395905
- PMCID: PMC6687829
- DOI: 10.1038/s41598-019-47795-0
Comparison of Two-Talker Attention Decoding from EEG with Nonlinear Neural Networks and Linear Methods
Abstract
Auditory attention decoding (AAD) through a brain-computer interface has had a flowering of developments since it was first introduced by Mesgarani and Chang (2012) using electrocorticograph recordings. AAD has been pursued for its potential application to hearing-aid design in which an attention-guided algorithm selects, from multiple competing acoustic sources, which should be enhanced for the listener and which should be suppressed. Traditionally, researchers have separated the AAD problem into two stages: reconstruction of a representation of the attended audio from neural signals, followed by determining the similarity between the candidate audio streams and the reconstruction. Here, we compare the traditional two-stage approach with a novel neural-network architecture that subsumes the explicit similarity step. We compare this new architecture against linear and non-linear (neural-network) baselines using both wet and dry electroencephalogram (EEG) systems. Our results indicate that the new architecture outperforms the baseline linear stimulus-reconstruction method, improving decoding accuracy from 66% to 81% using wet EEG and from 59% to 87% for dry EEG. Also of note was the finding that the dry EEG system can deliver comparable or even better results than the wet, despite the latter having one third as many EEG channels as the former. The 11-subject, wet-electrode AAD dataset for two competing, co-located talkers, the 11-subject, dry-electrode AAD dataset, and our software are available for further validation, experimentation, and modification.
Conflict of interest statement
All authors are part of a provisional patent application on the end-to-end, deep neural network auditory attention decoding algorithm described in this work. NM and J O’S are inventors on submitted patent WO2017218492A1 which covers neural decoding of auditory attention.
Figures
Similar articles
-
Noise-robust cortical tracking of attended speech in real-world acoustic scenes.Neuroimage. 2017 Aug 1;156:435-444. doi: 10.1016/j.neuroimage.2017.04.026. Epub 2017 Apr 13. Neuroimage. 2017. PMID: 28412441
-
Robust decoding of the speech envelope from EEG recordings through deep neural networks.J Neural Eng. 2022 Jul 6;19(4). doi: 10.1088/1741-2552/ac7976. J Neural Eng. 2022. PMID: 35709698
-
Congruent audiovisual speech enhances auditory attention decoding with EEG.J Neural Eng. 2019 Nov 6;16(6):066033. doi: 10.1088/1741-2552/ab4340. J Neural Eng. 2019. PMID: 31505476
-
Neural Encoding of Attended Continuous Speech under Different Types of Interference.J Cogn Neurosci. 2018 Nov;30(11):1606-1619. doi: 10.1162/jocn_a_01303. Epub 2018 Jul 13. J Cogn Neurosci. 2018. PMID: 30004849 Review.
-
Relating EEG to continuous speech using deep neural networks: a review.J Neural Eng. 2023 Aug 3;20(4). doi: 10.1088/1741-2552/ace73f. J Neural Eng. 2023. PMID: 37442115 Review.
Cited by
-
Distinct roles of SNR, speech Intelligibility, and attentional effort on neural speech tracking in noise.bioRxiv [Preprint]. 2024 Oct 12:2024.10.10.616515. doi: 10.1101/2024.10.10.616515. bioRxiv. 2024. PMID: 39416110 Free PMC article. Preprint.
-
A Tutorial on Auditory Attention Identification Methods.Front Neurosci. 2019 Mar 19;13:153. doi: 10.3389/fnins.2019.00153. eCollection 2019. Front Neurosci. 2019. PMID: 30941002 Free PMC article.
-
A GRU-CNN model for auditory attention detection using microstate and recurrence quantification analysis.Sci Rep. 2024 Apr 17;14(1):8861. doi: 10.1038/s41598-024-58886-y. Sci Rep. 2024. PMID: 38632246 Free PMC article.
-
Objective speech intelligibility prediction using a deep learning model with continuous speech-evoked cortical auditory responses.Front Neurosci. 2022 Aug 18;16:906616. doi: 10.3389/fnins.2022.906616. eCollection 2022. Front Neurosci. 2022. PMID: 36061597 Free PMC article.
-
Extracting the Auditory Attention in a Dual-Speaker Scenario From EEG Using a Joint CNN-LSTM Model.Front Physiol. 2021 Aug 2;12:700655. doi: 10.3389/fphys.2021.700655. eCollection 2021. Front Physiol. 2021. PMID: 34408661 Free PMC article.
References
-
- USVA. Annual Benefits Report Fiscal Year 2017. US Department of Veterans Affairs, Veterans Benefits Administration (2017).
-
- Kochkin S. Customer satisfaction with hearing instruments in the digital age. The Hearing Journal. 2005;58(9):30–43. doi: 10.1097/01.HJ.0000286545.33961.e7. - DOI
-
- Abrams H, Kihm J. An introduction to MarkeTrak IX: A new baseline for the hearing aid market. Hearing Review. 2015;22(6):16.
Publication types
MeSH terms
Grants and funding
LinkOut - more resources
Full Text Sources
Other Literature Sources
