Lip-reading aids word recognition most in moderate noise: a Bayesian explanation using high-dimensional feature space
- PMID: 19259259
- PMCID: PMC2645675
- DOI: 10.1371/journal.pone.0004638
Lip-reading aids word recognition most in moderate noise: a Bayesian explanation using high-dimensional feature space
Abstract
Watching a speaker's facial movements can dramatically enhance our ability to comprehend words, especially in noisy environments. From a general doctrine of combining information from different sensory modalities (the principle of inverse effectiveness), one would expect that the visual signals would be most effective at the highest levels of auditory noise. In contrast, we find, in accord with a recent paper, that visual information improves performance more at intermediate levels of auditory noise than at the highest levels, and we show that a novel visual stimulus containing only temporal information does the same. We present a Bayesian model of optimal cue integration that can explain these conflicts. In this model, words are regarded as points in a multidimensional space and word recognition is a probabilistic inference process. When the dimensionality of the feature space is low, the Bayesian model predicts inverse effectiveness; when the dimensionality is high, the enhancement is maximal at intermediate auditory noise levels. When the auditory and visual stimuli differ slightly in high noise, the model makes a counterintuitive prediction: as sound quality increases, the proportion of reported words corresponding to the visual stimulus should first increase and then decrease. We confirm this prediction in a behavioral experiment. We conclude that auditory-visual speech perception obeys the same notion of optimality previously observed only for simple multisensory stimuli.
Conflict of interest statement
Figures
Similar articles
-
The Principle of Inverse Effectiveness in Audiovisual Speech Perception.Front Hum Neurosci. 2019 Sep 26;13:335. doi: 10.3389/fnhum.2019.00335. eCollection 2019. Front Hum Neurosci. 2019. PMID: 31611780 Free PMC article.
-
Effects of audio-visual integration on the detection of masked speech and non-speech sounds.Brain Cogn. 2011 Feb;75(1):60-6. doi: 10.1016/j.bandc.2010.09.005. Epub 2010 Nov 9. Brain Cogn. 2011. PMID: 21067852
-
Aging, audiovisual integration, and the principle of inverse effectiveness.Ear Hear. 2010 Oct;31(5):636-44. doi: 10.1097/AUD.0b013e3181ddf7ff. Ear Hear. 2010. PMID: 20473178 Free PMC article.
-
Prediction and constraint in audiovisual speech perception.Cortex. 2015 Jul;68:169-81. doi: 10.1016/j.cortex.2015.03.006. Epub 2015 Mar 20. Cortex. 2015. PMID: 25890390 Free PMC article. Review.
-
A Multisensory Perspective on Human Auditory Communication.In: Murray MM, Wallace MT, editors. The Neural Bases of Multisensory Processes. Boca Raton (FL): CRC Press/Taylor & Francis; 2012. Chapter 34. In: Murray MM, Wallace MT, editors. The Neural Bases of Multisensory Processes. Boca Raton (FL): CRC Press/Taylor & Francis; 2012. Chapter 34. PMID: 22593871 Free Books & Documents. Review.
Cited by
-
The unity hypothesis revisited: can the male/female incongruent McGurk effect be disrupted by familiarization and priming?Front Psychol. 2023 Aug 29;14:1106562. doi: 10.3389/fpsyg.2023.1106562. eCollection 2023. Front Psychol. 2023. PMID: 37705948 Free PMC article.
-
Mouth and facial informativeness norms for 2276 English words.Behav Res Methods. 2023 Aug 21. doi: 10.3758/s13428-023-02216-z. Online ahead of print. Behav Res Methods. 2023. PMID: 37604959
-
Metacognition in the audiovisual McGurk illusion: perceptual and causal confidence.Philos Trans R Soc Lond B Biol Sci. 2023 Sep 25;378(1886):20220348. doi: 10.1098/rstb.2022.0348. Epub 2023 Aug 7. Philos Trans R Soc Lond B Biol Sci. 2023. PMID: 37545307 Free PMC article.
-
Benefit of visual speech information for word comprehension in post-stroke aphasia.Cortex. 2023 Aug;165:86-100. doi: 10.1016/j.cortex.2023.04.011. Epub 2023 May 16. Cortex. 2023. PMID: 37271014 Free PMC article.
-
Increases in sensory noise predict attentional disruptions to audiovisual speech perception.Front Hum Neurosci. 2023 Jan 4;16:1027335. doi: 10.3389/fnhum.2022.1027335. eCollection 2022. Front Hum Neurosci. 2023. PMID: 36684833 Free PMC article.
References
-
- Campbell R. Speechreading: advances in understanding its cortical bases and implications for deafness and speech rehabilitation. Scand Audiol. 1998;Suppl 49 - PubMed
-
- Bernstein LE, Demorest ME, Tucker PE. Speech perception without hearing. Perception and Psychophysics. Perception and Psychophysics. 2000;62:233–252. - PubMed
-
- Grant KW, Walden BE. Evaluating the articulation index for auditory-visual consonant recognition. J Acoust Soc Am. 1996;100:2415–2424. - PubMed
-
- MacLeod A, Summerfield Q. Quantifying the contribution of vision to speech perception in noise. Br J Audiol. 1987;21:131–141. - PubMed
-
- Massaro DW. Speech perception by ear and eye: A paradigm for psychological inquiry. Hillsdale, , NJ: Erlbaum; 1987.
MeSH terms
LinkOut - more resources
Full Text Sources
Other Literature Sources
Medical
