Inductive principles for learning Restricted Boltzmann Machines

We explore the training and usage of the Restricted Boltzmann Machine for unsupervised feature extraction. We investigate the many different aspects involved in their training, and by applying the concept of iterate averaging we show that it is possible to greatly improve on state of the art algorit...

Full description

Bibliographic Details
Main Author: Swersky, Kevin
Language:English
Published: University of British Columbia 2010
Online Access:http://hdl.handle.net/2429/27816
id ndltd-LACETR-oai-collectionscanada.gc.ca-BVAU.2429-27816
record_format oai_dc
spelling ndltd-LACETR-oai-collectionscanada.gc.ca-BVAU.2429-278162014-03-26T03:37:27Z Inductive principles for learning Restricted Boltzmann Machines Swersky, Kevin We explore the training and usage of the Restricted Boltzmann Machine for unsupervised feature extraction. We investigate the many different aspects involved in their training, and by applying the concept of iterate averaging we show that it is possible to greatly improve on state of the art algorithms. We also derive estimators based on the principles of pseudo-likelihood, ratio matching, and score matching, and we test them empirically against contrastive divergence, and stochastic maximum likelihood (also known as persistent contrastive divergence). Our results show that ratio matching and score matching are promising approaches to learning Restricted Boltzmann Machines. By applying score matching to the Restricted Boltzmann Machine, we show that training an auto-encoder neural network with a particular kind of regularization function is asymptotically consistent. Finally, we discuss the concept of deep learning and its relationship to training Restricted Boltzmann Machines, and briefly explore the impact of fine-tuning on the parameters and performance of a deep belief network. 2010-08-26T17:44:42Z 2010-08-26T17:44:42Z 2010 2010-08-26T17:44:42Z 2010-11 Electronic Thesis or Dissertation http://hdl.handle.net/2429/27816 eng University of British Columbia
collection NDLTD
language English
sources NDLTD
description We explore the training and usage of the Restricted Boltzmann Machine for unsupervised feature extraction. We investigate the many different aspects involved in their training, and by applying the concept of iterate averaging we show that it is possible to greatly improve on state of the art algorithms. We also derive estimators based on the principles of pseudo-likelihood, ratio matching, and score matching, and we test them empirically against contrastive divergence, and stochastic maximum likelihood (also known as persistent contrastive divergence). Our results show that ratio matching and score matching are promising approaches to learning Restricted Boltzmann Machines. By applying score matching to the Restricted Boltzmann Machine, we show that training an auto-encoder neural network with a particular kind of regularization function is asymptotically consistent. Finally, we discuss the concept of deep learning and its relationship to training Restricted Boltzmann Machines, and briefly explore the impact of fine-tuning on the parameters and performance of a deep belief network.
author Swersky, Kevin
spellingShingle Swersky, Kevin
Inductive principles for learning Restricted Boltzmann Machines
author_facet Swersky, Kevin
author_sort Swersky, Kevin
title Inductive principles for learning Restricted Boltzmann Machines
title_short Inductive principles for learning Restricted Boltzmann Machines
title_full Inductive principles for learning Restricted Boltzmann Machines
title_fullStr Inductive principles for learning Restricted Boltzmann Machines
title_full_unstemmed Inductive principles for learning Restricted Boltzmann Machines
title_sort inductive principles for learning restricted boltzmann machines
publisher University of British Columbia
publishDate 2010
url http://hdl.handle.net/2429/27816
work_keys_str_mv AT swerskykevin inductiveprinciplesforlearningrestrictedboltzmannmachines
_version_ 1716655695961522176