Learning matrix and functional models in high-dimensions

Statistical machine learning methods provide us with a principled framework for extracting meaningful information from noisy high-dimensional data sets. A significant feature of such procedures is that the inferences made are statistically significant, computationally efficient and scientifically me...

Full description

Bibliographic Details
Main Author: Balasubramanian, Krishnakumar
Other Authors: Lebanon, Guy
Format: Others
Language:en_US
Published: Georgia Institute of Technology 2014
Subjects:
Online Access:http://hdl.handle.net/1853/52284
id ndltd-GATECH-oai-smartech.gatech.edu-1853-52284
record_format oai_dc
spelling ndltd-GATECH-oai-smartech.gatech.edu-1853-522842014-09-13T03:33:56ZLearning matrix and functional models in high-dimensionsBalasubramanian, KrishnakumarStatisticsMachine learningMatrixKernelConsistencyStatistical machine learning methods provide us with a principled framework for extracting meaningful information from noisy high-dimensional data sets. A significant feature of such procedures is that the inferences made are statistically significant, computationally efficient and scientifically meaningful. In this thesis we make several contributions to such statistical procedures. Our contributions are two-fold. We first address prediction and estimation problems in non-standard situations. We show that even when given no access to labeled samples, one can still consistently estimate error rate of predictors and train predictors with respect to a given (convex) loss function. We next propose an efficient procedure for predicting with large output spaces, that scales logarithmically in the dimensionality of the output space. We further propose an asymptotically optimal procedure for sparse multi-task learning when the tasks share a joint support. We show consistency of the proposed method and derive rates of convergence. We next address the problem of learning meaningful representations of data. We propose a method for learning sparse representations that takes into account the structure of the data space and demonstrate how it enables one to obtain meaningful features. We establish sample complexity results for the proposed approach. We then propose a model-free feature selection procedure and establish its sure-screening property in the high dimensional regime. Furthermore we show that with a slight modification, the approach previously proposed for sparse multi-task learning enables one to obtain sparse representations for multiple related tasks simultaneously.Georgia Institute of TechnologyLebanon, Guy2014-08-27T13:38:53Z2014-08-27T13:38:53Z2014-082014-06-20August 20142014-08-27T13:38:53ZDissertationapplication/pdfhttp://hdl.handle.net/1853/52284en_US
collection NDLTD
language en_US
format Others
sources NDLTD
topic Statistics
Machine learning
Matrix
Kernel
Consistency
spellingShingle Statistics
Machine learning
Matrix
Kernel
Consistency
Balasubramanian, Krishnakumar
Learning matrix and functional models in high-dimensions
description Statistical machine learning methods provide us with a principled framework for extracting meaningful information from noisy high-dimensional data sets. A significant feature of such procedures is that the inferences made are statistically significant, computationally efficient and scientifically meaningful. In this thesis we make several contributions to such statistical procedures. Our contributions are two-fold. We first address prediction and estimation problems in non-standard situations. We show that even when given no access to labeled samples, one can still consistently estimate error rate of predictors and train predictors with respect to a given (convex) loss function. We next propose an efficient procedure for predicting with large output spaces, that scales logarithmically in the dimensionality of the output space. We further propose an asymptotically optimal procedure for sparse multi-task learning when the tasks share a joint support. We show consistency of the proposed method and derive rates of convergence. We next address the problem of learning meaningful representations of data. We propose a method for learning sparse representations that takes into account the structure of the data space and demonstrate how it enables one to obtain meaningful features. We establish sample complexity results for the proposed approach. We then propose a model-free feature selection procedure and establish its sure-screening property in the high dimensional regime. Furthermore we show that with a slight modification, the approach previously proposed for sparse multi-task learning enables one to obtain sparse representations for multiple related tasks simultaneously.
author2 Lebanon, Guy
author_facet Lebanon, Guy
Balasubramanian, Krishnakumar
author Balasubramanian, Krishnakumar
author_sort Balasubramanian, Krishnakumar
title Learning matrix and functional models in high-dimensions
title_short Learning matrix and functional models in high-dimensions
title_full Learning matrix and functional models in high-dimensions
title_fullStr Learning matrix and functional models in high-dimensions
title_full_unstemmed Learning matrix and functional models in high-dimensions
title_sort learning matrix and functional models in high-dimensions
publisher Georgia Institute of Technology
publishDate 2014
url http://hdl.handle.net/1853/52284
work_keys_str_mv AT balasubramaniankrishnakumar learningmatrixandfunctionalmodelsinhighdimensions
_version_ 1716714116142333952