Learning from other minds: An optimistic critique of reinforcement learning models of social learning

Curr Opin Behav Sci. 2021 Apr:38:110-115. doi: 10.1016/j.cobeha.2021.01.006. Epub 2021 Mar 23.

Abstract

Reinforcement learning models have been productively applied to identify neural correlates of the value of social information. However, by operationalizing social information as a lean, reward-predictive cue, this literature underestimates the richness of human social learning: Humans readily go beyond action-outcome mappings and can draw flexible inferences from a single observation. We argue that computational models of social learning need minds, i.e, a generative model of how others' unobservable mental states cause their observable actions. Recent advances in inferential social learning suggest that even young children learn from others by using an intuitive, generative model of other minds. Bridging developmental, Bayesian, and reinforcement learning perspectives can enrich our understanding of the neural bases of distinctively human social learning.

Keywords: Bayesian modeling; cognitive neuroscience; reinforcement learning; social learning.