Skip navigation
Please use this identifier to cite or link to this item: http://arks.princeton.edu/ark:/88435/dsp01fb494856q
Full metadata record
DC FieldValueLanguage
dc.contributor.advisorBotvinick, Matthew Men_US
dc.contributor.authorSolway, Alecen_US
dc.contributor.otherNeuroscience Departmenten_US
dc.date.accessioned2014-06-05T19:45:21Z-
dc.date.available2016-06-05T05:10:48Z-
dc.date.issued2014en_US
dc.identifier.urihttp://arks.princeton.edu/ark:/88435/dsp01fb494856q-
dc.description.abstractOver the last two decades, there has been a large scale effort in cognitive neuroscience to understand learning and decision making from the perspective of simple model-free reinforcement learning algorithms. This interest was invigorated in the mid 1990's, when it was realized that the phasic activity of midbrain dopaminergic neurons resembles reward prediction errors. The algorithms studied formalize the notion of learning from past experiences through trial and error. Although important, there are many aspects of behavior they cannot explain. More recent work has begun to fill in some of these gaps by borrowing yet additional ideas from computational reinforcement learning. One line of inquiry has concentrated on aligning goal-directed behavior, which resembles the common sense notion of &ldquo;planning&rdquo;, with <italic>model-based</italic> reinforcement learning. This work has aimed to understand how the brain is able to learn the world model prescribed by the model-based framework, and to characterize the neural correlates of the value functions it predicts. This thesis adds to this work by offering two separate, but related, algorithmic accounts of how the brain may be able to actually map the world model into a decision. Existing data are examined and new experiments are performed. A second line of inquiry has concentrated on understanding behavior from the perspective of <italic>hierarchical</italic> reinforcement learning. The thesis makes two contributions to this area as well. First, it is shown that the brain codes pseudo-reward prediction errors, a prediction error in response to a faux reward signal that is used to train skills that are not in themselves useful, but that may be used to achieve other means. Second, an optimality framework is provided for understanding which skills are most beneficial to have when confronted with an ensemble of tasks.en_US
dc.language.isoenen_US
dc.publisherPrinceton, NJ : Princeton Universityen_US
dc.relation.isformatofThe Mudd Manuscript Library retains one bound copy of each dissertation. Search for these copies in the <a href=http://catalog.princeton.edu> library's main catalog </a>en_US
dc.subject.classificationNeurosciencesen_US
dc.titleBeyond simple model-free reinforcement learning in human decision makingen_US
dc.typeAcademic dissertations (Ph.D.)en_US
pu.projectgrantnumber690-2143en_US
pu.embargo.terms2016-06-05en_US
Appears in Collections:Neuroscience

Files in This Item:
File Description SizeFormat 
Solway_princeton_0181D_10921.pdf17.23 MBAdobe PDFView/Download


Items in Dataspace are protected by copyright, with all rights reserved, unless otherwise indicated.