Skip navigation
Please use this identifier to cite or link to this item: http://arks.princeton.edu/ark:/88435/dsp01kp78gj709
Full metadata record
DC FieldValueLanguage
dc.contributor.advisorImai, Kosuke-
dc.contributor.authorAbdurehman, Rahji-
dc.date.accessioned2015-07-01T18:28:15Z-
dc.date.available2015-07-01T18:28:15Z-
dc.date.created2015-04-30-
dc.date.issued2015-07-01-
dc.identifier.urihttp://arks.princeton.edu/ark:/88435/dsp01kp78gj709-
dc.description.abstractThis paper introduces an alternative to the popular machine learning algorithm known as Latent Dirichlet Allocation, or LDA for short. In this paper we derive the theory behind this alternative algorithm and demonstrate a specific use case for it with sample results. We call this new algorithm "keyword-assisted LDA". It works by taking a set of constraints which are set based on prior knowledge of the underlying topic structure within a corpus and then ensuring that they are maintained. Depending on one’s underlying implementation of LDA, keeping these constraints in order takes a variety of forms. This paper delves into the details for implementations using Gibbs sampling or Expectation-Maximization.en_US
dc.format.extent36 pagesen_US
dc.language.isoen_USen_US
dc.titleKeyword-assisted LDA: Exploring New Methods for Supervised Topic Modelingen_US
dc.typePrinceton University Senior Theses-
pu.date.classyear2015en_US
pu.departmentComputer Scienceen_US
pu.pdf.coverpageSeniorThesisCoverPage-
dc.rights.accessRightsWalk-in Access. This thesis can only be viewed on computer terminals at the <a href=http://mudd.princeton.edu>Mudd Manuscript Library</a>.*
Appears in Collections:Computer Science, 1987-2023

Files in This Item:
File SizeFormat 
PUTheses2015-Abdurehman_Rahji.pdf609.88 kBAdobe PDF    Request a copy


Items in Dataspace are protected by copyright, with all rights reserved, unless otherwise indicated.