Maximum entropy methods for extracting the learned features of deep neural networks.

New architectures of multilayer artificial neural networks and new methods for training them are rapidly revolutionizing the application of machine learning in diverse fields, including business, social science, physical sciences, and biology. Interpreting deep neural networks, however, currently re...

Full description

Bibliographic Details
Main Authors: Alex Finnegan, Jun S Song
Format: Article
Language:English
Published: Public Library of Science (PLoS) 2017-10-01
Series:PLoS Computational Biology
Online Access:http://europepmc.org/articles/PMC5679649?pdf=render
id doaj-eb1f1c63c2a64ec5875d292787713da7
record_format Article
spelling doaj-eb1f1c63c2a64ec5875d292787713da72020-11-25T01:37:03ZengPublic Library of Science (PLoS)PLoS Computational Biology1553-734X1553-73582017-10-011310e100583610.1371/journal.pcbi.1005836Maximum entropy methods for extracting the learned features of deep neural networks.Alex FinneganJun S SongNew architectures of multilayer artificial neural networks and new methods for training them are rapidly revolutionizing the application of machine learning in diverse fields, including business, social science, physical sciences, and biology. Interpreting deep neural networks, however, currently remains elusive, and a critical challenge lies in understanding which meaningful features a network is actually learning. We present a general method for interpreting deep neural networks and extracting network-learned features from input data. We describe our algorithm in the context of biological sequence analysis. Our approach, based on ideas from statistical physics, samples from the maximum entropy distribution over possible sequences, anchored at an input sequence and subject to constraints implied by the empirical function learned by a network. Using our framework, we demonstrate that local transcription factor binding motifs can be identified from a network trained on ChIP-seq data and that nucleosome positioning signals are indeed learned by a network trained on chemical cleavage nucleosome maps. Imposing a further constraint on the maximum entropy distribution also allows us to probe whether a network is learning global sequence features, such as the high GC content in nucleosome-rich regions. This work thus provides valuable mathematical tools for interpreting and extracting learned features from feed-forward neural networks.http://europepmc.org/articles/PMC5679649?pdf=render
collection DOAJ
language English
format Article
sources DOAJ
author Alex Finnegan
Jun S Song
spellingShingle Alex Finnegan
Jun S Song
Maximum entropy methods for extracting the learned features of deep neural networks.
PLoS Computational Biology
author_facet Alex Finnegan
Jun S Song
author_sort Alex Finnegan
title Maximum entropy methods for extracting the learned features of deep neural networks.
title_short Maximum entropy methods for extracting the learned features of deep neural networks.
title_full Maximum entropy methods for extracting the learned features of deep neural networks.
title_fullStr Maximum entropy methods for extracting the learned features of deep neural networks.
title_full_unstemmed Maximum entropy methods for extracting the learned features of deep neural networks.
title_sort maximum entropy methods for extracting the learned features of deep neural networks.
publisher Public Library of Science (PLoS)
series PLoS Computational Biology
issn 1553-734X
1553-7358
publishDate 2017-10-01
description New architectures of multilayer artificial neural networks and new methods for training them are rapidly revolutionizing the application of machine learning in diverse fields, including business, social science, physical sciences, and biology. Interpreting deep neural networks, however, currently remains elusive, and a critical challenge lies in understanding which meaningful features a network is actually learning. We present a general method for interpreting deep neural networks and extracting network-learned features from input data. We describe our algorithm in the context of biological sequence analysis. Our approach, based on ideas from statistical physics, samples from the maximum entropy distribution over possible sequences, anchored at an input sequence and subject to constraints implied by the empirical function learned by a network. Using our framework, we demonstrate that local transcription factor binding motifs can be identified from a network trained on ChIP-seq data and that nucleosome positioning signals are indeed learned by a network trained on chemical cleavage nucleosome maps. Imposing a further constraint on the maximum entropy distribution also allows us to probe whether a network is learning global sequence features, such as the high GC content in nucleosome-rich regions. This work thus provides valuable mathematical tools for interpreting and extracting learned features from feed-forward neural networks.
url http://europepmc.org/articles/PMC5679649?pdf=render
work_keys_str_mv AT alexfinnegan maximumentropymethodsforextractingthelearnedfeaturesofdeepneuralnetworks
AT junssong maximumentropymethodsforextractingthelearnedfeaturesofdeepneuralnetworks
_version_ 1725059900721070080