Please use this identifier to cite or link to this item:

`http://arks.princeton.edu/ark:/88435/dsp01b8515q61f`

Full metadata record

DC Field | Value | Language |
---|---|---|

dc.contributor.advisor | Charikar, Moses S | en_US |

dc.contributor.author | Nguyen, Huy Le | en_US |

dc.contributor.other | Computer Science Department | en_US |

dc.date.accessioned | 2014-09-25T22:42:13Z | - |

dc.date.available | 2014-09-25T22:42:13Z | - |

dc.date.issued | 2014 | en_US |

dc.identifier.uri | http://arks.princeton.edu/ark:/88435/dsp01b8515q61f | - |

dc.description.abstract | The quest of understanding large high dimension data has always been plagued by the so-called "curse of dimensionality": algorithms regularly suffer from terrible exponential dependence on the dimension. One popular approach to deal with this problem is <italic>dimension reduction</italic>, the process of reducing high dimensional objects to lower dimensional objects or other compact representations while maintaining the objectives of the problems at hand. This philosophy manifests itself in many different forms: from analytic notions like <italic>metric embeddings</italic>, to <italic>locality sensitive hashing</italic>, and even information theoretic notions like <italic>sketches</italic>. A fundamental task in data processing is identifying patterns, from similar objects, to relations between different attributes of the same object. An abstract way to model similarity is via the notion of <italic>distance</italic>: two objects are similar if their distance is small, a relation of regressors is predictive if the observed dependent variable is at a small distance from the predicted value, etc. All three aforementioned notions preserve distances but with varied amounts of structure. Metric embeddings preserve the most amount of structure and hence, can be combined easily with existing algorithms. In contrast, sketches require no additional structure and therefore, are the easiest to obtain but often require new algorithms. Locality sensitive hashing only has limited structure, which is enough to enable certain applications such as nearest neighbor search. The flexibility of such a large toolkit is hugely beneficial. We will demonstrate applications of these techniques to design fast algorithms for practical questions including nearest neighbor search, linear regressions, and low rank approximation, study their limitations, and overcome those limitations with new techniques. Specifically, in this thesis, we present the following contributions. <italic>Fast dimension reduction for sparse data.</italic> In many applications, the input data is in high dimensions but also sparse. Examples of such inputs include columns of document-word incidence matrices in information retrieval, the person-object scoring matrix in collaborative filtering tasks, constraint matrices for geodetic survey problems, photogrammetry problems, molecular structure problems [Rice83], etc. In the last decade, subspace embedding has emerged as a useful tool for designing algorithms for such data. We present a new construction of subspace embedding with nearly optimal dimensions and embedding time nearly linear in the sparsity of the input. Our new embeddings have improved the running time of many important problems from least squares regression to low rank approximations, which are workhorses of statistical analysis. For instance, our running time for least squares regression is nearly optimal unless one can avoid generic matrix multiplication. <italic>Lower bounds for dimension reduction.</italic> We show the embeddings above are in fact nearly optimal in many regimes of parameters. Additionally, the same techniques also yield nearly tight bounds for sparse Johnson-Lindenstrass embedding, and RIP matrices. We also give new lower bounds for dimension reduction in L1, showing that for 1+ε distortion, there are point sets of size n in L1 requiring n^(1-1/log (1/ε)) dimensions. This bound is nearly optimal as one can reduce the dimensions to linear while approximately preserving all distances. <italic>Better approximate nearest neighbor search.</italic> For approximation factor $c$, we give a new algorithm for approximate nearest neighbor search in Euclidean space with query time n^ρ and space n^(1+ρ) where ρ = 7/(8c^2) + O(1/c^3). This algorithm bypasses the limit of locality sensitive hashing, the main previous technique for approximate nearest neighbor search in high dimensions. Furthermore, the algorithm achieves faster running time and smaller space via data-dependent hashing, a method widely used in practice but eluded satisfactory mathematical explanation. | en_US |

dc.language.iso | en | en_US |

dc.publisher | Princeton, NJ : Princeton University | en_US |

dc.relation.isformatof | The Mudd Manuscript Library retains one bound copy of each dissertation. Search for these copies in the <a href=http://catalog.princeton.edu> library's main catalog </a> | en_US |

dc.subject | algorithms | en_US |

dc.subject | dimension reduction | en_US |

dc.subject | high dimension | en_US |

dc.subject | nearest neighbor search | en_US |

dc.subject | regression | en_US |

dc.subject.classification | Computer science | en_US |

dc.title | Algorithms for High Dimensional Data | en_US |

dc.type | Academic dissertations (Ph.D.) | en_US |

pu.projectgrantnumber | 690-2143 | en_US |

Appears in Collections: | Computer Science |

Files in This Item:

File | Description | Size | Format | |
---|---|---|---|---|

Nguyen_princeton_0181D_11063.pdf | 1.32 MB | Adobe PDF | View/Download |

Items in Dataspace are protected by copyright, with all rights reserved, unless otherwise indicated.