Recklessly approximate sparse coding

Introduction of the so called “K-means” or “triangle” features in Coates, Lee and Ng, 2011 caused significant discussion in the deep learning community. These simple features are able to achieve state of the art performance on standard image classification benchmarks, outperforming much more soph...

Full description

Bibliographic Details
Main Author: Denil, Misha
Language:English
Published: University of British Columbia 2012
Online Access:http://hdl.handle.net/2429/43662
id ndltd-LACETR-oai-collectionscanada.gc.ca-BVAU.2429-43662
record_format oai_dc
spelling ndltd-LACETR-oai-collectionscanada.gc.ca-BVAU.2429-436622014-03-26T03:39:11Z Recklessly approximate sparse coding Denil, Misha Introduction of the so called “K-means” or “triangle” features in Coates, Lee and Ng, 2011 caused significant discussion in the deep learning community. These simple features are able to achieve state of the art performance on standard image classification benchmarks, outperforming much more sophisticated methods including deep belief networks, convolutional nets, factored RBMs, mcRBMs, convolutional RBMs, sparse autoencoders and several others. Moreover, these features are extremely simple and easy to compute. Several intuitive arguments have been put forward to describe this remarkable performance, yet no mathematical justification has been offered. In Coates and Ng, 2011, the authors improve on the triangle features with “soft threshold” features, adding a hyperparameter to tune performance, and compare these features to sparse coding. Both soft thresholding and sparse coding are found to often yield similar classification results, though soft threshold features are much faster to compute. The main result of this thesis is to show that the soft threshold features are realized as a single step of proximal gradient descent on a non-negative sparse coding objective. This result is important because it provides an explanation for the success of the soft threshold features and shows that even very approximate solutions to the sparse coding problem are sufficient to build effective classifiers. 2012-12-06T19:30:58Z 2012-12-06T19:30:58Z 2012 2012-12-06 2013-05 Electronic Thesis or Dissertation http://hdl.handle.net/2429/43662 eng University of British Columbia
collection NDLTD
language English
sources NDLTD
description Introduction of the so called “K-means” or “triangle” features in Coates, Lee and Ng, 2011 caused significant discussion in the deep learning community. These simple features are able to achieve state of the art performance on standard image classification benchmarks, outperforming much more sophisticated methods including deep belief networks, convolutional nets, factored RBMs, mcRBMs, convolutional RBMs, sparse autoencoders and several others. Moreover, these features are extremely simple and easy to compute. Several intuitive arguments have been put forward to describe this remarkable performance, yet no mathematical justification has been offered. In Coates and Ng, 2011, the authors improve on the triangle features with “soft threshold” features, adding a hyperparameter to tune performance, and compare these features to sparse coding. Both soft thresholding and sparse coding are found to often yield similar classification results, though soft threshold features are much faster to compute. The main result of this thesis is to show that the soft threshold features are realized as a single step of proximal gradient descent on a non-negative sparse coding objective. This result is important because it provides an explanation for the success of the soft threshold features and shows that even very approximate solutions to the sparse coding problem are sufficient to build effective classifiers.
author Denil, Misha
spellingShingle Denil, Misha
Recklessly approximate sparse coding
author_facet Denil, Misha
author_sort Denil, Misha
title Recklessly approximate sparse coding
title_short Recklessly approximate sparse coding
title_full Recklessly approximate sparse coding
title_fullStr Recklessly approximate sparse coding
title_full_unstemmed Recklessly approximate sparse coding
title_sort recklessly approximate sparse coding
publisher University of British Columbia
publishDate 2012
url http://hdl.handle.net/2429/43662
work_keys_str_mv AT denilmisha recklesslyapproximatesparsecoding
_version_ 1716656550905380864