Recklessly approximate sparse coding
Introduction of the so called “K-means” or “triangle” features in Coates, Lee and Ng, 2011 caused significant discussion in the deep learning community. These simple features are able to achieve state of the art performance on standard image classification benchmarks, outperforming much more soph...
Main Author: | |
---|---|
Language: | English |
Published: |
University of British Columbia
2012
|
Online Access: | http://hdl.handle.net/2429/43662 |
id |
ndltd-LACETR-oai-collectionscanada.gc.ca-BVAU.2429-43662 |
---|---|
record_format |
oai_dc |
spelling |
ndltd-LACETR-oai-collectionscanada.gc.ca-BVAU.2429-436622014-03-26T03:39:11Z Recklessly approximate sparse coding Denil, Misha Introduction of the so called “K-means” or “triangle” features in Coates, Lee and Ng, 2011 caused significant discussion in the deep learning community. These simple features are able to achieve state of the art performance on standard image classification benchmarks, outperforming much more sophisticated methods including deep belief networks, convolutional nets, factored RBMs, mcRBMs, convolutional RBMs, sparse autoencoders and several others. Moreover, these features are extremely simple and easy to compute. Several intuitive arguments have been put forward to describe this remarkable performance, yet no mathematical justification has been offered. In Coates and Ng, 2011, the authors improve on the triangle features with “soft threshold” features, adding a hyperparameter to tune performance, and compare these features to sparse coding. Both soft thresholding and sparse coding are found to often yield similar classification results, though soft threshold features are much faster to compute. The main result of this thesis is to show that the soft threshold features are realized as a single step of proximal gradient descent on a non-negative sparse coding objective. This result is important because it provides an explanation for the success of the soft threshold features and shows that even very approximate solutions to the sparse coding problem are sufficient to build effective classifiers. 2012-12-06T19:30:58Z 2012-12-06T19:30:58Z 2012 2012-12-06 2013-05 Electronic Thesis or Dissertation http://hdl.handle.net/2429/43662 eng University of British Columbia |
collection |
NDLTD |
language |
English |
sources |
NDLTD |
description |
Introduction of the so called “K-means” or “triangle” features in Coates,
Lee and Ng, 2011 caused significant discussion in the deep learning
community. These simple features are able to achieve state of the art performance on standard image classification benchmarks, outperforming much
more sophisticated methods including deep belief networks, convolutional
nets, factored RBMs, mcRBMs, convolutional RBMs, sparse autoencoders
and several others. Moreover, these features are extremely simple and easy
to compute.
Several intuitive arguments have been put forward to describe this remarkable performance, yet no mathematical justification has been offered.
In Coates and Ng, 2011, the authors improve on the triangle features
with “soft threshold” features, adding a hyperparameter to tune performance, and compare these features to sparse coding. Both soft thresholding and sparse coding are found to often yield similar classification results,
though soft threshold features are much faster to compute.
The main result of this thesis is to show that the soft threshold features
are realized as a single step of proximal gradient descent on a non-negative
sparse coding objective. This result is important because it provides an
explanation for the success of the soft threshold features and shows that
even very approximate solutions to the sparse coding problem are sufficient
to build effective classifiers. |
author |
Denil, Misha |
spellingShingle |
Denil, Misha Recklessly approximate sparse coding |
author_facet |
Denil, Misha |
author_sort |
Denil, Misha |
title |
Recklessly approximate sparse coding |
title_short |
Recklessly approximate sparse coding |
title_full |
Recklessly approximate sparse coding |
title_fullStr |
Recklessly approximate sparse coding |
title_full_unstemmed |
Recklessly approximate sparse coding |
title_sort |
recklessly approximate sparse coding |
publisher |
University of British Columbia |
publishDate |
2012 |
url |
http://hdl.handle.net/2429/43662 |
work_keys_str_mv |
AT denilmisha recklesslyapproximatesparsecoding |
_version_ |
1716656550905380864 |