High-frequency neural activity predicts word parsing in ambiguous speech streams

J Neurophysiol. 2016 Dec 1;116(6):2497-2512. doi: 10.1152/jn.00074.2016. Epub 2016 Sep 7.


During speech listening, the brain parses a continuous acoustic stream of information into computational units (e.g., syllables or words) necessary for speech comprehension. Recent neuroscientific hypotheses have proposed that neural oscillations contribute to speech parsing, but whether they do so on the basis of acoustic cues (bottom-up acoustic parsing) or as a function of available linguistic representations (top-down linguistic parsing) is unknown. In this magnetoencephalography study, we contrasted acoustic and linguistic parsing using bistable speech sequences. While listening to the speech sequences, participants were asked to maintain one of the two possible speech percepts through volitional control. We predicted that the tracking of speech dynamics by neural oscillations would not only follow the acoustic properties but also shift in time according to the participant's conscious speech percept. Our results show that the latency of high-frequency activity (specifically, beta and gamma bands) varied as a function of the perceptual report. In contrast, the phase of low-frequency oscillations was not strongly affected by top-down control. Whereas changes in low-frequency neural oscillations were compatible with the encoding of prelexical segmentation cues, high-frequency activity specifically informed on an individual's conscious speech percept.

Keywords: MEG; bistability; neural entrainment; phase; speech segmentation.

Publication types

  • Research Support, Non-U.S. Gov't

MeSH terms

  • Acoustic Stimulation
  • Analysis of Variance
  • Brain Mapping*
  • Brain Waves / physiology*
  • Comprehension / physiology*
  • Female
  • Humans
  • Linguistics
  • Magnetoencephalography
  • Male
  • Reaction Time / physiology
  • Spectrum Analysis
  • Speech / physiology*
  • Speech Perception / physiology*
  • Vocabulary
  • Young Adult