Scaling prediction errors to reward variability benefits error-driven learning in humans

J Neurophysiol. 2015 Sep;114(3):1628-40. doi: 10.1152/jn.00483.2015. Epub 2015 Jul 15.

Abstract

Effective error-driven learning requires individuals to adapt learning to environmental reward variability. The adaptive mechanism may involve decays in learning rate across subsequent trials, as shown previously, and rescaling of reward prediction errors. The present study investigated the influence of prediction error scaling and, in particular, the consequences for learning performance. Participants explicitly predicted reward magnitudes that were drawn from different probability distributions with specific standard deviations. By fitting the data with reinforcement learning models, we found scaling of prediction errors, in addition to the learning rate decay shown previously. Importantly, the prediction error scaling was closely related to learning performance, defined as accuracy in predicting the mean of reward distributions, across individual participants. In addition, participants who scaled prediction errors relative to standard deviation also presented with more similar performance for different standard deviations, indicating that increases in standard deviation did not substantially decrease "adapters'" accuracy in predicting the means of reward distributions. However, exaggerated scaling beyond the standard deviation resulted in impaired performance. Thus efficient adaptation makes learning more robust to changing variability.

Keywords: adaptation; probability distribution; reinforcement learning; risk; standard deviation.

Publication types

  • Research Support, Non-U.S. Gov't

MeSH terms

  • Adolescent
  • Adult
  • Anticipation, Psychological
  • Female
  • Humans
  • Knowledge of Results, Psychological*
  • Male
  • Models, Neurological
  • Models, Statistical
  • Reward*