Intrinsic Gradient Networks

<p>Artificial neural networks are computationally powerful and exhibit brain-like dynamics. Unfortunately, the conventional gradient-dependent learning algorithms used to train them are biologically implausible. The calculation of the gradient in a traditional artificial neural network require...

Full description

Bibliographic Details
Main Author: Rolfe, Jason Tyler
Format: Others
Published: 2012
Online Access:https://thesis.library.caltech.edu/6953/1/rolfe_jason_2012_thesis.pdf
Rolfe, Jason Tyler (2012) Intrinsic Gradient Networks. Dissertation (Ph.D.), California Institute of Technology. doi:10.7907/YCB7-7X24. https://resolver.caltech.edu/CaltechTHESIS:04202012-133210844 <https://resolver.caltech.edu/CaltechTHESIS:04202012-133210844>
id ndltd-CALTECH-oai-thesis.library.caltech.edu-6953
record_format oai_dc
spelling ndltd-CALTECH-oai-thesis.library.caltech.edu-69532019-10-04T03:12:02Z Intrinsic Gradient Networks Rolfe, Jason Tyler <p>Artificial neural networks are computationally powerful and exhibit brain-like dynamics. Unfortunately, the conventional gradient-dependent learning algorithms used to train them are biologically implausible. The calculation of the gradient in a traditional artificial neural network requires a complementary network of fast training signals that are dependent upon, but must not affect, the primary output-generating network activity. In contrast, the network of neurons in the cortex is highly recurrent; a network of gradient-calculating neurons in the brain would certainly project back to and influence the primary network. We address this biological implausibility by introducing a novel class of recurrent neural networks, intrinsic gradient networks, for which the gradient of an error function with respect to the parameters is a simple function of the network state. These networks can be trained using only their intrinsic signals, much like the network of neurons in the brain.</p> <p>We derive a simple equation that characterizes intrinsic gradient networks, and construct a broad set of networks that satisfy this characteristic equation. The resulting set of intrinsic gradient networks includes many highly recurrent instances for which the gradient can be calculated by a simple, local, pseudo-Hebbian function of the network state, thus resolving a long-standing contradiction between artificial and biological neural networks. We demonstrate that these networks can learn to perform nontrivial tasks like handwritten digit recognition using only their intrinsic signals. Finally, we show that a cortical implementation of an intrinsic gradient network would have a number of characteristic computational, anatomical, and electrophysiological properties, and review experimental evidence suggesting the manifestation of these properties in the cortex.</p> 2012 Thesis NonPeerReviewed application/pdf https://thesis.library.caltech.edu/6953/1/rolfe_jason_2012_thesis.pdf https://resolver.caltech.edu/CaltechTHESIS:04202012-133210844 Rolfe, Jason Tyler (2012) Intrinsic Gradient Networks. Dissertation (Ph.D.), California Institute of Technology. doi:10.7907/YCB7-7X24. https://resolver.caltech.edu/CaltechTHESIS:04202012-133210844 <https://resolver.caltech.edu/CaltechTHESIS:04202012-133210844> https://thesis.library.caltech.edu/6953/
collection NDLTD
format Others
sources NDLTD
description <p>Artificial neural networks are computationally powerful and exhibit brain-like dynamics. Unfortunately, the conventional gradient-dependent learning algorithms used to train them are biologically implausible. The calculation of the gradient in a traditional artificial neural network requires a complementary network of fast training signals that are dependent upon, but must not affect, the primary output-generating network activity. In contrast, the network of neurons in the cortex is highly recurrent; a network of gradient-calculating neurons in the brain would certainly project back to and influence the primary network. We address this biological implausibility by introducing a novel class of recurrent neural networks, intrinsic gradient networks, for which the gradient of an error function with respect to the parameters is a simple function of the network state. These networks can be trained using only their intrinsic signals, much like the network of neurons in the brain.</p> <p>We derive a simple equation that characterizes intrinsic gradient networks, and construct a broad set of networks that satisfy this characteristic equation. The resulting set of intrinsic gradient networks includes many highly recurrent instances for which the gradient can be calculated by a simple, local, pseudo-Hebbian function of the network state, thus resolving a long-standing contradiction between artificial and biological neural networks. We demonstrate that these networks can learn to perform nontrivial tasks like handwritten digit recognition using only their intrinsic signals. Finally, we show that a cortical implementation of an intrinsic gradient network would have a number of characteristic computational, anatomical, and electrophysiological properties, and review experimental evidence suggesting the manifestation of these properties in the cortex.</p>
author Rolfe, Jason Tyler
spellingShingle Rolfe, Jason Tyler
Intrinsic Gradient Networks
author_facet Rolfe, Jason Tyler
author_sort Rolfe, Jason Tyler
title Intrinsic Gradient Networks
title_short Intrinsic Gradient Networks
title_full Intrinsic Gradient Networks
title_fullStr Intrinsic Gradient Networks
title_full_unstemmed Intrinsic Gradient Networks
title_sort intrinsic gradient networks
publishDate 2012
url https://thesis.library.caltech.edu/6953/1/rolfe_jason_2012_thesis.pdf
Rolfe, Jason Tyler (2012) Intrinsic Gradient Networks. Dissertation (Ph.D.), California Institute of Technology. doi:10.7907/YCB7-7X24. https://resolver.caltech.edu/CaltechTHESIS:04202012-133210844 <https://resolver.caltech.edu/CaltechTHESIS:04202012-133210844>
work_keys_str_mv AT rolfejasontyler intrinsicgradientnetworks
_version_ 1719259675289976832