Please use this identifier to cite or link to this item:
|Title:||Tensor Decomposition and Memory Networks for SAT Reading Comprehension|
|Abstract:||Abstract We present a general approach for machine comprehension tasks by converting the text to a knowledge graph and the questions to queries on the graph. We extend  and use the Stanford NLP Toolkit’s Dependency Parser [17, 9] to transform each sentence into a set of entity-relation triples. We use word2vec  to convert the questions into queries on the graph. We present a tensor decomposition approach to answering queries by adding Semantically Smooth Embedding  to RESCAL . We also generalize the Memory Networks [28, 25] architecture to take any knowledge graph as input. We evaluate these models on three full SAT reading comprehension tests. The models presented here outperform their respective baselines. Both models demonstrate the ability to capture the semantic and structural information in the text and answer questions using that information.|
|Type of Material:||Princeton University Senior Theses|
|Appears in Collections:||Computer Science, 1988-2016|
Files in This Item:
|Madge_Saahil_2016_thesis.pdf||905.23 kB||Adobe PDF||Request a copy|
Items in Dataspace are protected by copyright, with all rights reserved, unless otherwise indicated.