Semantic Coherence Facilitates Distributional Learning

Cogn Sci. 2017 Apr;41 Suppl 4:855-884. doi: 10.1111/cogs.12360. Epub 2016 Mar 14.


Computational models have shown that purely statistical knowledge about words' linguistic contexts is sufficient to learn many properties of words, including syntactic and semantic category. For example, models can infer that "postman" and "mailman" are semantically similar because they have quantitatively similar patterns of association with other words (e.g., they both tend to occur with words like "deliver," "truck," "package"). In contrast to these computational results, artificial language learning experiments suggest that distributional statistics alone do not facilitate learning of linguistic categories. However, experiments in this paradigm expose participants to entirely novel words, whereas real language learners encounter input that contains some known words that are semantically organized. In three experiments, we show that (a) the presence of familiar semantic reference points facilitates distributional learning and (b) this effect crucially depends both on the presence of known words and the adherence of these known words to some semantic organization.

Keywords: Distributional learning; Semantic coherence; Word learning.

MeSH terms

  • Cues
  • Humans
  • Language Development*
  • Language*
  • Learning*
  • Verbal Learning*
  • Vocabulary*