Information Theory for Analyzing Neural Networks

The goal of this thesis was to investigate how information theory could be used to analyze artificial neural networks. For this purpose, two problems, a classification problem and a controller problem were considered. The classification problem was solved with a feedforward neural network trained wi...

Full description

Bibliographic Details
Main Author: Sørngård, Bård
Format: Others
Language:English
Published: Norges teknisk-naturvitenskapelige universitet, Institutt for datateknikk og informasjonsvitenskap 2014
Online Access:http://urn.kb.se/resolve?urn=urn:nbn:no:ntnu:diva-26773
id ndltd-UPSALLA1-oai-DiVA.org-ntnu-26773
record_format oai_dc
spelling ndltd-UPSALLA1-oai-DiVA.org-ntnu-267732014-10-01T04:57:18ZInformation Theory for Analyzing Neural NetworksengSørngård, BårdNorges teknisk-naturvitenskapelige universitet, Institutt for datateknikk og informasjonsvitenskapInstitutt for datateknikk og informasjonsvitenskap2014The goal of this thesis was to investigate how information theory could be used to analyze artificial neural networks. For this purpose, two problems, a classification problem and a controller problem were considered. The classification problem was solved with a feedforward neural network trained with backpropagation, the controller problem was solved with a continuous-time recurrent neural network optimized with evolution.Results from the classification problem shows that mutual information might indicate how much a particular neuron contributes to the classification. Tracking these neurons' mutual information during training might serve as an indicator of their progression, including neurons in the hidden layers.Results from the controller problem showed that time-delayed mutual information between a neuron and an environment variable might indicate what variable each neuron is estimating, and tracking this during evolution might tell us when this particular neuron started taking this role. Furthermore, unrolled transfer entropy appears to be a good measure for how neurons affect each other during simulation. Student thesisinfo:eu-repo/semantics/bachelorThesistexthttp://urn.kb.se/resolve?urn=urn:nbn:no:ntnu:diva-26773Local ntnudaim:11773application/pdfinfo:eu-repo/semantics/openAccess
collection NDLTD
language English
format Others
sources NDLTD
description The goal of this thesis was to investigate how information theory could be used to analyze artificial neural networks. For this purpose, two problems, a classification problem and a controller problem were considered. The classification problem was solved with a feedforward neural network trained with backpropagation, the controller problem was solved with a continuous-time recurrent neural network optimized with evolution.Results from the classification problem shows that mutual information might indicate how much a particular neuron contributes to the classification. Tracking these neurons' mutual information during training might serve as an indicator of their progression, including neurons in the hidden layers.Results from the controller problem showed that time-delayed mutual information between a neuron and an environment variable might indicate what variable each neuron is estimating, and tracking this during evolution might tell us when this particular neuron started taking this role. Furthermore, unrolled transfer entropy appears to be a good measure for how neurons affect each other during simulation.
author Sørngård, Bård
spellingShingle Sørngård, Bård
Information Theory for Analyzing Neural Networks
author_facet Sørngård, Bård
author_sort Sørngård, Bård
title Information Theory for Analyzing Neural Networks
title_short Information Theory for Analyzing Neural Networks
title_full Information Theory for Analyzing Neural Networks
title_fullStr Information Theory for Analyzing Neural Networks
title_full_unstemmed Information Theory for Analyzing Neural Networks
title_sort information theory for analyzing neural networks
publisher Norges teknisk-naturvitenskapelige universitet, Institutt for datateknikk og informasjonsvitenskap
publishDate 2014
url http://urn.kb.se/resolve?urn=urn:nbn:no:ntnu:diva-26773
work_keys_str_mv AT sørngardbard informationtheoryforanalyzingneuralnetworks
_version_ 1716715992993759232