Learning matrix and functional models in high-dimensions
Statistical machine learning methods provide us with a principled framework for extracting meaningful information from noisy high-dimensional data sets. A significant feature of such procedures is that the inferences made are statistically significant, computationally efficient and scientifically me...
Main Author: | |
---|---|
Other Authors: | |
Format: | Others |
Language: | en_US |
Published: |
Georgia Institute of Technology
2014
|
Subjects: | |
Online Access: | http://hdl.handle.net/1853/52284 |
id |
ndltd-GATECH-oai-smartech.gatech.edu-1853-52284 |
---|---|
record_format |
oai_dc |
spelling |
ndltd-GATECH-oai-smartech.gatech.edu-1853-522842014-09-13T03:33:56ZLearning matrix and functional models in high-dimensionsBalasubramanian, KrishnakumarStatisticsMachine learningMatrixKernelConsistencyStatistical machine learning methods provide us with a principled framework for extracting meaningful information from noisy high-dimensional data sets. A significant feature of such procedures is that the inferences made are statistically significant, computationally efficient and scientifically meaningful. In this thesis we make several contributions to such statistical procedures. Our contributions are two-fold. We first address prediction and estimation problems in non-standard situations. We show that even when given no access to labeled samples, one can still consistently estimate error rate of predictors and train predictors with respect to a given (convex) loss function. We next propose an efficient procedure for predicting with large output spaces, that scales logarithmically in the dimensionality of the output space. We further propose an asymptotically optimal procedure for sparse multi-task learning when the tasks share a joint support. We show consistency of the proposed method and derive rates of convergence. We next address the problem of learning meaningful representations of data. We propose a method for learning sparse representations that takes into account the structure of the data space and demonstrate how it enables one to obtain meaningful features. We establish sample complexity results for the proposed approach. We then propose a model-free feature selection procedure and establish its sure-screening property in the high dimensional regime. Furthermore we show that with a slight modification, the approach previously proposed for sparse multi-task learning enables one to obtain sparse representations for multiple related tasks simultaneously.Georgia Institute of TechnologyLebanon, Guy2014-08-27T13:38:53Z2014-08-27T13:38:53Z2014-082014-06-20August 20142014-08-27T13:38:53ZDissertationapplication/pdfhttp://hdl.handle.net/1853/52284en_US |
collection |
NDLTD |
language |
en_US |
format |
Others
|
sources |
NDLTD |
topic |
Statistics Machine learning Matrix Kernel Consistency |
spellingShingle |
Statistics Machine learning Matrix Kernel Consistency Balasubramanian, Krishnakumar Learning matrix and functional models in high-dimensions |
description |
Statistical machine learning methods provide us with a principled framework for extracting meaningful information from noisy high-dimensional data sets. A significant feature of such procedures is that the inferences made are statistically significant, computationally efficient and scientifically meaningful. In this thesis we make several contributions to such statistical procedures. Our contributions are two-fold.
We first address prediction and estimation problems in non-standard situations. We show that even when given no access to labeled samples, one can still consistently estimate error rate of predictors and train predictors with respect to a given (convex) loss function. We next propose an efficient procedure for predicting with large output spaces, that scales logarithmically in the dimensionality of the output space. We further propose an asymptotically optimal procedure for sparse multi-task learning when the tasks share a joint support. We show consistency of the proposed method and derive rates of convergence.
We next address the problem of learning meaningful representations of data. We propose a method for learning sparse representations that takes into account the structure of the data space and demonstrate how it enables one to obtain meaningful features. We establish sample complexity results for the proposed approach. We then propose a model-free feature selection procedure and establish its sure-screening property in the high dimensional regime. Furthermore we show that with a slight modification, the approach previously proposed for sparse multi-task learning enables one to obtain sparse representations for multiple related tasks simultaneously. |
author2 |
Lebanon, Guy |
author_facet |
Lebanon, Guy Balasubramanian, Krishnakumar |
author |
Balasubramanian, Krishnakumar |
author_sort |
Balasubramanian, Krishnakumar |
title |
Learning matrix and functional models in high-dimensions |
title_short |
Learning matrix and functional models in high-dimensions |
title_full |
Learning matrix and functional models in high-dimensions |
title_fullStr |
Learning matrix and functional models in high-dimensions |
title_full_unstemmed |
Learning matrix and functional models in high-dimensions |
title_sort |
learning matrix and functional models in high-dimensions |
publisher |
Georgia Institute of Technology |
publishDate |
2014 |
url |
http://hdl.handle.net/1853/52284 |
work_keys_str_mv |
AT balasubramaniankrishnakumar learningmatrixandfunctionalmodelsinhighdimensions |
_version_ |
1716714116142333952 |