On Pixel-Wise Explanations for Non-Linear Classifier Decisions by Layer-Wise Relevance Propagation.

Understanding and interpreting classification decisions of automated image classification systems is of high value in many applications, as it allows to verify the reasoning of the system and provides additional information to the human expert. Although machine learning methods are solving very succ...

Full description

Bibliographic Details
Main Authors: Sebastian Bach, Alexander Binder, Grégoire Montavon, Frederick Klauschen, Klaus-Robert Müller, Wojciech Samek
Format: Article
Language:English
Published: Public Library of Science (PLoS) 2015-01-01
Series:PLoS ONE
Online Access:http://europepmc.org/articles/PMC4498753?pdf=render
id doaj-1889e731d1be48abb891b7c2d7c4c899
record_format Article
spelling doaj-1889e731d1be48abb891b7c2d7c4c8992020-11-25T01:24:21ZengPublic Library of Science (PLoS)PLoS ONE1932-62032015-01-01107e013014010.1371/journal.pone.0130140On Pixel-Wise Explanations for Non-Linear Classifier Decisions by Layer-Wise Relevance Propagation.Sebastian BachAlexander BinderGrégoire MontavonFrederick KlauschenKlaus-Robert MüllerWojciech SamekUnderstanding and interpreting classification decisions of automated image classification systems is of high value in many applications, as it allows to verify the reasoning of the system and provides additional information to the human expert. Although machine learning methods are solving very successfully a plethora of tasks, they have in most cases the disadvantage of acting as a black box, not providing any information about what made them arrive at a particular decision. This work proposes a general solution to the problem of understanding classification decisions by pixel-wise decomposition of nonlinear classifiers. We introduce a methodology that allows to visualize the contributions of single pixels to predictions for kernel-based classifiers over Bag of Words features and for multilayered neural networks. These pixel contributions can be visualized as heatmaps and are provided to a human expert who can intuitively not only verify the validity of the classification decision, but also focus further analysis on regions of potential interest. We evaluate our method for classifiers trained on PASCAL VOC 2009 images, synthetic image data containing geometric shapes, the MNIST handwritten digits data set and for the pre-trained ImageNet model available as part of the Caffe open source package.http://europepmc.org/articles/PMC4498753?pdf=render
collection DOAJ
language English
format Article
sources DOAJ
author Sebastian Bach
Alexander Binder
Grégoire Montavon
Frederick Klauschen
Klaus-Robert Müller
Wojciech Samek
spellingShingle Sebastian Bach
Alexander Binder
Grégoire Montavon
Frederick Klauschen
Klaus-Robert Müller
Wojciech Samek
On Pixel-Wise Explanations for Non-Linear Classifier Decisions by Layer-Wise Relevance Propagation.
PLoS ONE
author_facet Sebastian Bach
Alexander Binder
Grégoire Montavon
Frederick Klauschen
Klaus-Robert Müller
Wojciech Samek
author_sort Sebastian Bach
title On Pixel-Wise Explanations for Non-Linear Classifier Decisions by Layer-Wise Relevance Propagation.
title_short On Pixel-Wise Explanations for Non-Linear Classifier Decisions by Layer-Wise Relevance Propagation.
title_full On Pixel-Wise Explanations for Non-Linear Classifier Decisions by Layer-Wise Relevance Propagation.
title_fullStr On Pixel-Wise Explanations for Non-Linear Classifier Decisions by Layer-Wise Relevance Propagation.
title_full_unstemmed On Pixel-Wise Explanations for Non-Linear Classifier Decisions by Layer-Wise Relevance Propagation.
title_sort on pixel-wise explanations for non-linear classifier decisions by layer-wise relevance propagation.
publisher Public Library of Science (PLoS)
series PLoS ONE
issn 1932-6203
publishDate 2015-01-01
description Understanding and interpreting classification decisions of automated image classification systems is of high value in many applications, as it allows to verify the reasoning of the system and provides additional information to the human expert. Although machine learning methods are solving very successfully a plethora of tasks, they have in most cases the disadvantage of acting as a black box, not providing any information about what made them arrive at a particular decision. This work proposes a general solution to the problem of understanding classification decisions by pixel-wise decomposition of nonlinear classifiers. We introduce a methodology that allows to visualize the contributions of single pixels to predictions for kernel-based classifiers over Bag of Words features and for multilayered neural networks. These pixel contributions can be visualized as heatmaps and are provided to a human expert who can intuitively not only verify the validity of the classification decision, but also focus further analysis on regions of potential interest. We evaluate our method for classifiers trained on PASCAL VOC 2009 images, synthetic image data containing geometric shapes, the MNIST handwritten digits data set and for the pre-trained ImageNet model available as part of the Caffe open source package.
url http://europepmc.org/articles/PMC4498753?pdf=render
work_keys_str_mv AT sebastianbach onpixelwiseexplanationsfornonlinearclassifierdecisionsbylayerwiserelevancepropagation
AT alexanderbinder onpixelwiseexplanationsfornonlinearclassifierdecisionsbylayerwiserelevancepropagation
AT gregoiremontavon onpixelwiseexplanationsfornonlinearclassifierdecisionsbylayerwiserelevancepropagation
AT frederickklauschen onpixelwiseexplanationsfornonlinearclassifierdecisionsbylayerwiserelevancepropagation
AT klausrobertmuller onpixelwiseexplanationsfornonlinearclassifierdecisionsbylayerwiserelevancepropagation
AT wojciechsamek onpixelwiseexplanationsfornonlinearclassifierdecisionsbylayerwiserelevancepropagation
_version_ 1725117678970994688