Please use this identifier to cite or link to this item: http://arks.princeton.edu/ark:/88435/dsp011r66j348x
 Title: Asymmetric Learning Rates for Positive and Negative Feedback: A Formal Model Comparison Authors: McDonald, Kelsey Advisors: Niv, Yael Contributors: Norman, Ken Department: Psychology Class Year: 2015 Abstract: Reinforcement learning studies how learning systems interact with their environments in order to maximize a numerical reward signal. One of the central concepts in reinforcement learning is the reward “prediction error”, which is the numerical value of the reward received minus the expected reward value. The classic Rescorla-Wagner model posits that the learning agent updates their original reward estimate by stepwise error-correction: multiplying the prediction error by a learning rate parameter. The main limitation with the Rescorla-Wagner model is the implied valence symmetry with which feedback updates an action’s value estimate. This contradicts evidence that learning from positive and negative feedback has different effects on behavior and decision-making. In this thesis, I conduct a formal model comparison of the Rescorla-Wagner model with an alternative class of asymmetric learning models which discriminate based on valence. Our analyses show that behavioral choice data in a probabilistic learning experiment is more accurately described by an asymmetric learning algorithm rather than a symmetric learning rule which does not discriminate based on valence. Extent: 63 pages URI: http://arks.princeton.edu/ark:/88435/dsp011r66j348x Type of Material: Princeton University Senior Theses Language: en_US Appears in Collections: Psychology, 1930-2016

Files in This Item:
File SizeFormat