Modality-constrained statistical learning of tactile, visual, and auditory sequences

J Exp Psychol Learn Mem Cogn. 2005 Jan;31(1):24-39. doi: 10.1037/0278-7393.31.1.24.

Abstract

The authors investigated the extent to which touch, vision, and audition mediate the processing of statistical regularities within sequential input. Few researchers have conducted rigorous comparisons across sensory modalities; in particular, the sense of touch has been virtually ignored. The current data reveal not only commonalities but also modality constraints affecting statistical learning across the senses. To be specific, the authors found that the auditory modality displayed a quantitative learning advantage compared with vision and touch. In addition, they discovered qualitative learning biases among the senses: Primarily, audition afforded better learning for the final part of input sequences. These findings are discussed in terms of whether statistical learning is likely to consist of a single, unitary mechanism or multiple, modality-constrained ones.

Publication types

  • Clinical Trial
  • Randomized Controlled Trial

MeSH terms

  • Auditory Perception*
  • Humans
  • Learning*
  • Psychology / statistics & numerical data*
  • Reaction Time
  • Time Perception*
  • Touch*
  • Visual Perception*