Explainable AI Metrics and Properties for Evaluation and Analysis of Counterfactual Explanations : Explainable AI Metrics and Properties for Evaluation and Analysis of Counterfactual Explanations
With the recent surge of state-of-the-art AI systems, there has been a growing need to provide explanations for these systems to ensure transparency and understandability by the stakeholders. This has directed the development of increasingly popular Explainable AI systems applicable to a wide range...
Main Author: | |
---|---|
Format: | Others |
Language: | English |
Published: |
Uppsala universitet, Institutionen för informationsteknologi
2021
|
Subjects: | |
Online Access: | http://urn.kb.se/resolve?urn=urn:nbn:se:uu:diva-462552 |
id |
ndltd-UPSALLA1-oai-DiVA.org-uu-462552 |
---|---|
record_format |
oai_dc |
spelling |
ndltd-UPSALLA1-oai-DiVA.org-uu-4625522021-12-28T06:00:00ZExplainable AI Metrics and Properties for Evaluation and Analysis of Counterfactual Explanations : Explainable AI Metrics and Properties for Evaluation and Analysis of Counterfactual ExplanationsengSingh, VanditaUppsala universitet, Institutionen för informationsteknologi2021Engineering and TechnologyTeknik och teknologierWith the recent surge of state-of-the-art AI systems, there has been a growing need to provide explanations for these systems to ensure transparency and understandability by the stakeholders. This has directed the development of increasingly popular Explainable AI systems applicable to a wide range of applications. This trend designates the need for the ability to quantitatively assess and analyze the theoretical and behavioral characteristics of these Explainable AI systems. This study aimed at the following (i) to identify metrics and properties applicable to post-hoc counterfactual explanation methods (a mechanism for generating explanations), (ii) assess the applicability of the identified metrics and properties to compare counterfactual examples, and (iii) analyze the functional and operational characteristics of the selected counterfactual explanation methods. A proof-of-concept tool was designed to perform the given task. A pipeline was implemented, which comprised of selecting a data set, training some suitable classifier, deploying counterfactual generation method, and performing analysis using the metrics selected for implementation. Four counterfactual explanation methods were selected and analyzed against the metrics - distance measures, loss measures, change score values, recourse values in terms of the value of features, and neighborhood. It was found that the metrics provided a means to analyze the characteristics of the selected counterfactual explanation methods. It could also be reasoned which explainers generated better counterfactual explanations for a particular instance by comparing the metrics across different explainers for the same instance. Also, insight could be gained to understand whether a particular counterfactual instance was useful enough or not by comparing the implemented metrics concerning an instance of the desired class. A fact sheet enlisting the functional and operational characteristics of the selected counterfactual methods has been presented. The scope of the work was limited to data sets being comprised of tabular data and images, simpler classification methods were chosen to be incorporated into the pipeline, and the type of explanations was confined to counterfactual explanations. Student thesisinfo:eu-repo/semantics/bachelorThesistexthttp://urn.kb.se/resolve?urn=urn:nbn:se:uu:diva-462552IT ; 21120application/pdfinfo:eu-repo/semantics/openAccess |
collection |
NDLTD |
language |
English |
format |
Others
|
sources |
NDLTD |
topic |
Engineering and Technology Teknik och teknologier |
spellingShingle |
Engineering and Technology Teknik och teknologier Singh, Vandita Explainable AI Metrics and Properties for Evaluation and Analysis of Counterfactual Explanations : Explainable AI Metrics and Properties for Evaluation and Analysis of Counterfactual Explanations |
description |
With the recent surge of state-of-the-art AI systems, there has been a growing need to provide explanations for these systems to ensure transparency and understandability by the stakeholders. This has directed the development of increasingly popular Explainable AI systems applicable to a wide range of applications. This trend designates the need for the ability to quantitatively assess and analyze the theoretical and behavioral characteristics of these Explainable AI systems. This study aimed at the following (i) to identify metrics and properties applicable to post-hoc counterfactual explanation methods (a mechanism for generating explanations), (ii) assess the applicability of the identified metrics and properties to compare counterfactual examples, and (iii) analyze the functional and operational characteristics of the selected counterfactual explanation methods. A proof-of-concept tool was designed to perform the given task. A pipeline was implemented, which comprised of selecting a data set, training some suitable classifier, deploying counterfactual generation method, and performing analysis using the metrics selected for implementation. Four counterfactual explanation methods were selected and analyzed against the metrics - distance measures, loss measures, change score values, recourse values in terms of the value of features, and neighborhood. It was found that the metrics provided a means to analyze the characteristics of the selected counterfactual explanation methods. It could also be reasoned which explainers generated better counterfactual explanations for a particular instance by comparing the metrics across different explainers for the same instance. Also, insight could be gained to understand whether a particular counterfactual instance was useful enough or not by comparing the implemented metrics concerning an instance of the desired class. A fact sheet enlisting the functional and operational characteristics of the selected counterfactual methods has been presented. The scope of the work was limited to data sets being comprised of tabular data and images, simpler classification methods were chosen to be incorporated into the pipeline, and the type of explanations was confined to counterfactual explanations. |
author |
Singh, Vandita |
author_facet |
Singh, Vandita |
author_sort |
Singh, Vandita |
title |
Explainable AI Metrics and Properties for Evaluation and Analysis of Counterfactual Explanations : Explainable AI Metrics and Properties for Evaluation and Analysis of Counterfactual Explanations |
title_short |
Explainable AI Metrics and Properties for Evaluation and Analysis of Counterfactual Explanations : Explainable AI Metrics and Properties for Evaluation and Analysis of Counterfactual Explanations |
title_full |
Explainable AI Metrics and Properties for Evaluation and Analysis of Counterfactual Explanations : Explainable AI Metrics and Properties for Evaluation and Analysis of Counterfactual Explanations |
title_fullStr |
Explainable AI Metrics and Properties for Evaluation and Analysis of Counterfactual Explanations : Explainable AI Metrics and Properties for Evaluation and Analysis of Counterfactual Explanations |
title_full_unstemmed |
Explainable AI Metrics and Properties for Evaluation and Analysis of Counterfactual Explanations : Explainable AI Metrics and Properties for Evaluation and Analysis of Counterfactual Explanations |
title_sort |
explainable ai metrics and properties for evaluation and analysis of counterfactual explanations : explainable ai metrics and properties for evaluation and analysis of counterfactual explanations |
publisher |
Uppsala universitet, Institutionen för informationsteknologi |
publishDate |
2021 |
url |
http://urn.kb.se/resolve?urn=urn:nbn:se:uu:diva-462552 |
work_keys_str_mv |
AT singhvandita explainableaimetricsandpropertiesforevaluationandanalysisofcounterfactualexplanationsexplainableaimetricsandpropertiesforevaluationandanalysisofcounterfactualexplanations |
_version_ |
1723965802311319552 |