Concentration of measure, negative association, and machine learning
In this thesis we consider concentration inequalities and the concentration of measure phenomenon from a variety of angles. Sharp tail bounds on the deviation of Lipschitz functions of independent random variables about their mean are well known. We consider variations on this theme for dependent v...
Main Author: | |
---|---|
Language: | en_US |
Published: |
2016
|
Subjects: | |
Online Access: | https://hdl.handle.net/2144/19741 |
id |
ndltd-bu.edu-oai-open.bu.edu-2144-19741 |
---|---|
record_format |
oai_dc |
spelling |
ndltd-bu.edu-oai-open.bu.edu-2144-197412019-01-08T15:40:41Z Concentration of measure, negative association, and machine learning Root, Jonathan Mathematics Anomaly detection Compressed sensing Concentration inequalities Concentration of measure Machine learning Negative association In this thesis we consider concentration inequalities and the concentration of measure phenomenon from a variety of angles. Sharp tail bounds on the deviation of Lipschitz functions of independent random variables about their mean are well known. We consider variations on this theme for dependent variables on the Boolean cube. In recent years negatively associated probability distributions have been studied as potential generalizations of independent random variables. Results on this class of distributions have been sparse at best, even when restricting to the Boolean cube. We consider the class of negatively associated distributions topologically, as a subset of the general class of probability measures. Both the weak (distributional) topology and the total variation topology are considered, and the simpler notion of negative correlation is investigated. The concentration of measure phenomenon began with Milman's proof of Dvoretzky's theorem, and is therefore intimately connected to the field of high-dimensional convex geometry. Recently this field has found application in the area of compressed sensing. We consider these applications and in particular analyze the use of Gordon's min-max inequality in various compressed sensing frameworks, including the Dantzig selector and the matrix uncertainty selector. Finally we consider the use of concentration inequalities in developing a theoretically sound anomaly detection algorithm. Our method uses a ranking procedure based on KNN graphs of given data. We develop a max-margin learning-to-rank framework to train limited complexity models to imitate these KNN scores. The resulting anomaly detector is shown to be asymptotically optimal in that for any false alarm rate α, its decision region converges to the α-percentile minimum volume level set of the unknown underlying density. 2016-12-20T19:58:47Z 2016-12-20T19:58:47Z 2016 2016-12-07T02:08:05Z Thesis/Dissertation https://hdl.handle.net/2144/19741 en_US |
collection |
NDLTD |
language |
en_US |
sources |
NDLTD |
topic |
Mathematics Anomaly detection Compressed sensing Concentration inequalities Concentration of measure Machine learning Negative association |
spellingShingle |
Mathematics Anomaly detection Compressed sensing Concentration inequalities Concentration of measure Machine learning Negative association Root, Jonathan Concentration of measure, negative association, and machine learning |
description |
In this thesis we consider concentration inequalities and the concentration
of measure phenomenon from a variety of angles. Sharp tail bounds on the deviation of Lipschitz functions of independent random variables about their mean are well known. We consider variations on this theme for dependent variables on the Boolean cube.
In recent years negatively associated probability distributions have been studied as potential generalizations of independent random variables. Results on this class of distributions have been sparse at best, even when restricting to the Boolean cube. We consider the class of negatively associated distributions topologically, as a subset of the general class of probability measures. Both the weak (distributional) topology and the total variation topology
are considered, and the simpler notion of negative correlation is investigated.
The concentration of measure phenomenon began with Milman's proof of Dvoretzky's theorem, and is therefore intimately connected to the field of high-dimensional convex geometry. Recently this field has found application in the area of compressed sensing. We consider these applications and in particular analyze the use of Gordon's min-max inequality in various compressed sensing frameworks, including the Dantzig selector and the matrix uncertainty selector.
Finally we consider the use of concentration inequalities in developing a theoretically sound anomaly detection algorithm. Our method uses a ranking procedure based on KNN graphs
of given data. We develop a max-margin learning-to-rank framework to train limited complexity models to imitate these KNN scores. The resulting anomaly detector is shown to be asymptotically optimal in that for any false alarm rate α, its decision region converges to the α-percentile minimum volume level set of the unknown
underlying density. |
author |
Root, Jonathan |
author_facet |
Root, Jonathan |
author_sort |
Root, Jonathan |
title |
Concentration of measure, negative association, and machine learning |
title_short |
Concentration of measure, negative association, and machine learning |
title_full |
Concentration of measure, negative association, and machine learning |
title_fullStr |
Concentration of measure, negative association, and machine learning |
title_full_unstemmed |
Concentration of measure, negative association, and machine learning |
title_sort |
concentration of measure, negative association, and machine learning |
publishDate |
2016 |
url |
https://hdl.handle.net/2144/19741 |
work_keys_str_mv |
AT rootjonathan concentrationofmeasurenegativeassociationandmachinelearning |
_version_ |
1718812036786290688 |