Dimensionality reduction for sparse and structured matrices

Thesis: S.M., Massachusetts Institute of Technology, Department of Electrical Engineering and Computer Science, 2015. === Cataloged from PDF version of thesis. === Includes bibliographical references (pages 97-103). === Dimensionality reduction has become a critical tool for quickly solving massive...

Full description

Bibliographic Details
Main Author: Musco, Christopher Paul
Other Authors: Martin C. Rinard and Jonathan A. Kelner.
Format: Others
Language:English
Published: Massachusetts Institute of Technology 2015
Subjects:
Online Access:http://hdl.handle.net/1721.1/99856
id ndltd-MIT-oai-dspace.mit.edu-1721.1-99856
record_format oai_dc
spelling ndltd-MIT-oai-dspace.mit.edu-1721.1-998562019-05-02T15:52:42Z Dimensionality reduction for sparse and structured matrices Musco, Christopher Paul Martin C. Rinard and Jonathan A. Kelner. Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science. Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science. Electrical Engineering and Computer Science. Thesis: S.M., Massachusetts Institute of Technology, Department of Electrical Engineering and Computer Science, 2015. Cataloged from PDF version of thesis. Includes bibliographical references (pages 97-103). Dimensionality reduction has become a critical tool for quickly solving massive matrix problems. Especially in modern data analysis and machine learning applications, an overabundance of data features or examples can make it impossible to apply standard algorithms efficiently. To address this issue, it is often possible to distill data to a much smaller set of informative features or examples, which can be used to obtain provably accurate approximate solutions to a variety of problems In this thesis, we focus on the important case of dimensionality reduction for sparse and structured data. In contrast to popular structure-agnostic methods like Johnson-Lindenstrauss projection and PCA, we seek data compression techniques that take advantage of structure to generate smaller or more powerful compressions. Additionally, we aim for methods that can be applied extremely quickly - typically in linear or nearly linear time in the input size. Specifically, we introduce new randomized algorithms for structured dimensionality reduction that are based on importance sampling and sparse-recovery techniques. Our work applies directly to accelerating linear regression and graph sparsification and we discuss connections and possible extensions to low-rank approximation, k-means clustering, and several other ubiquitous matrix problems. by Christopher Paul Musco. S.M. 2015-11-09T19:53:22Z 2015-11-09T19:53:22Z 2015 2015 Thesis http://hdl.handle.net/1721.1/99856 927699160 eng M.I.T. theses are protected by copyright. They may be viewed from this source for any purpose, but reproduction or distribution in any format is prohibited without written permission. See provided URL for inquiries about permission. http://dspace.mit.edu/handle/1721.1/7582 103 pages application/pdf Massachusetts Institute of Technology
collection NDLTD
language English
format Others
sources NDLTD
topic Electrical Engineering and Computer Science.
spellingShingle Electrical Engineering and Computer Science.
Musco, Christopher Paul
Dimensionality reduction for sparse and structured matrices
description Thesis: S.M., Massachusetts Institute of Technology, Department of Electrical Engineering and Computer Science, 2015. === Cataloged from PDF version of thesis. === Includes bibliographical references (pages 97-103). === Dimensionality reduction has become a critical tool for quickly solving massive matrix problems. Especially in modern data analysis and machine learning applications, an overabundance of data features or examples can make it impossible to apply standard algorithms efficiently. To address this issue, it is often possible to distill data to a much smaller set of informative features or examples, which can be used to obtain provably accurate approximate solutions to a variety of problems In this thesis, we focus on the important case of dimensionality reduction for sparse and structured data. In contrast to popular structure-agnostic methods like Johnson-Lindenstrauss projection and PCA, we seek data compression techniques that take advantage of structure to generate smaller or more powerful compressions. Additionally, we aim for methods that can be applied extremely quickly - typically in linear or nearly linear time in the input size. Specifically, we introduce new randomized algorithms for structured dimensionality reduction that are based on importance sampling and sparse-recovery techniques. Our work applies directly to accelerating linear regression and graph sparsification and we discuss connections and possible extensions to low-rank approximation, k-means clustering, and several other ubiquitous matrix problems. === by Christopher Paul Musco. === S.M.
author2 Martin C. Rinard and Jonathan A. Kelner.
author_facet Martin C. Rinard and Jonathan A. Kelner.
Musco, Christopher Paul
author Musco, Christopher Paul
author_sort Musco, Christopher Paul
title Dimensionality reduction for sparse and structured matrices
title_short Dimensionality reduction for sparse and structured matrices
title_full Dimensionality reduction for sparse and structured matrices
title_fullStr Dimensionality reduction for sparse and structured matrices
title_full_unstemmed Dimensionality reduction for sparse and structured matrices
title_sort dimensionality reduction for sparse and structured matrices
publisher Massachusetts Institute of Technology
publishDate 2015
url http://hdl.handle.net/1721.1/99856
work_keys_str_mv AT muscochristopherpaul dimensionalityreductionforsparseandstructuredmatrices
_version_ 1719030372792008704