Information Theory for Analyzing Neural Networks
The goal of this thesis was to investigate how information theory could be used to analyze artificial neural networks. For this purpose, two problems, a classification problem and a controller problem were considered. The classification problem was solved with a feedforward neural network trained wi...
Main Author: | |
---|---|
Format: | Others |
Language: | English |
Published: |
Norges teknisk-naturvitenskapelige universitet, Institutt for datateknikk og informasjonsvitenskap
2014
|
Online Access: | http://urn.kb.se/resolve?urn=urn:nbn:no:ntnu:diva-26773 |
id |
ndltd-UPSALLA1-oai-DiVA.org-ntnu-26773 |
---|---|
record_format |
oai_dc |
spelling |
ndltd-UPSALLA1-oai-DiVA.org-ntnu-267732014-10-01T04:57:18ZInformation Theory for Analyzing Neural NetworksengSørngård, BårdNorges teknisk-naturvitenskapelige universitet, Institutt for datateknikk og informasjonsvitenskapInstitutt for datateknikk og informasjonsvitenskap2014The goal of this thesis was to investigate how information theory could be used to analyze artificial neural networks. For this purpose, two problems, a classification problem and a controller problem were considered. The classification problem was solved with a feedforward neural network trained with backpropagation, the controller problem was solved with a continuous-time recurrent neural network optimized with evolution.Results from the classification problem shows that mutual information might indicate how much a particular neuron contributes to the classification. Tracking these neurons' mutual information during training might serve as an indicator of their progression, including neurons in the hidden layers.Results from the controller problem showed that time-delayed mutual information between a neuron and an environment variable might indicate what variable each neuron is estimating, and tracking this during evolution might tell us when this particular neuron started taking this role. Furthermore, unrolled transfer entropy appears to be a good measure for how neurons affect each other during simulation. Student thesisinfo:eu-repo/semantics/bachelorThesistexthttp://urn.kb.se/resolve?urn=urn:nbn:no:ntnu:diva-26773Local ntnudaim:11773application/pdfinfo:eu-repo/semantics/openAccess |
collection |
NDLTD |
language |
English |
format |
Others
|
sources |
NDLTD |
description |
The goal of this thesis was to investigate how information theory could be used to analyze artificial neural networks. For this purpose, two problems, a classification problem and a controller problem were considered. The classification problem was solved with a feedforward neural network trained with backpropagation, the controller problem was solved with a continuous-time recurrent neural network optimized with evolution.Results from the classification problem shows that mutual information might indicate how much a particular neuron contributes to the classification. Tracking these neurons' mutual information during training might serve as an indicator of their progression, including neurons in the hidden layers.Results from the controller problem showed that time-delayed mutual information between a neuron and an environment variable might indicate what variable each neuron is estimating, and tracking this during evolution might tell us when this particular neuron started taking this role. Furthermore, unrolled transfer entropy appears to be a good measure for how neurons affect each other during simulation. |
author |
Sørngård, Bård |
spellingShingle |
Sørngård, Bård Information Theory for Analyzing Neural Networks |
author_facet |
Sørngård, Bård |
author_sort |
Sørngård, Bård |
title |
Information Theory for Analyzing Neural Networks |
title_short |
Information Theory for Analyzing Neural Networks |
title_full |
Information Theory for Analyzing Neural Networks |
title_fullStr |
Information Theory for Analyzing Neural Networks |
title_full_unstemmed |
Information Theory for Analyzing Neural Networks |
title_sort |
information theory for analyzing neural networks |
publisher |
Norges teknisk-naturvitenskapelige universitet, Institutt for datateknikk og informasjonsvitenskap |
publishDate |
2014 |
url |
http://urn.kb.se/resolve?urn=urn:nbn:no:ntnu:diva-26773 |
work_keys_str_mv |
AT sørngardbard informationtheoryforanalyzingneuralnetworks |
_version_ |
1716715992993759232 |