Explainable AI Metrics and Properties for Evaluation and Analysis of Counterfactual Explanations : Explainable AI Metrics and Properties for Evaluation and Analysis of Counterfactual Explanations

With the recent surge of state-of-the-art AI systems, there has been a growing need to provide explanations for these systems to ensure transparency and understandability by the stakeholders. This has directed the development of increasingly popular Explainable AI systems applicable to a wide range...

Full description

Bibliographic Details
Main Author: Singh, Vandita
Format: Others
Language:English
Published: Uppsala universitet, Institutionen för informationsteknologi 2021
Subjects:
Online Access:http://urn.kb.se/resolve?urn=urn:nbn:se:uu:diva-462552
id ndltd-UPSALLA1-oai-DiVA.org-uu-462552
record_format oai_dc
spelling ndltd-UPSALLA1-oai-DiVA.org-uu-4625522021-12-28T06:00:00ZExplainable AI Metrics and Properties for Evaluation and Analysis of Counterfactual Explanations : Explainable AI Metrics and Properties for Evaluation and Analysis of Counterfactual ExplanationsengSingh, VanditaUppsala universitet, Institutionen för informationsteknologi2021Engineering and TechnologyTeknik och teknologierWith the recent surge of state-of-the-art AI systems, there has been a growing need to provide explanations for these systems to ensure transparency and understandability by the stakeholders. This has directed the development of increasingly popular Explainable AI systems applicable to a wide range of applications. This trend designates the need for the ability to quantitatively assess and analyze the theoretical and behavioral characteristics of these Explainable AI systems. This study aimed at the following (i) to identify metrics and properties applicable to post-hoc counterfactual explanation methods (a mechanism for generating explanations), (ii) assess the applicability of the identified metrics and properties to compare counterfactual examples, and (iii) analyze the functional and operational characteristics of the selected counterfactual explanation methods. A proof-of-concept tool was designed to perform the given task. A pipeline was implemented, which comprised of selecting a data set, training some suitable classifier, deploying counterfactual generation method, and performing analysis using the metrics selected for implementation. Four counterfactual explanation methods were selected and analyzed against the metrics - distance measures, loss measures, change score values, recourse values in terms of the value of features, and neighborhood. It was found that the metrics provided a means to analyze the characteristics of the selected counterfactual explanation methods. It could also be reasoned which explainers generated better counterfactual explanations for a particular instance by comparing the metrics across different explainers for the same instance. Also, insight could be gained to understand whether a particular counterfactual instance was useful enough or not by comparing the implemented metrics concerning an instance of the desired class. A fact sheet enlisting the functional and operational characteristics of the selected counterfactual methods has been presented. The scope of the work was limited to data sets being comprised of tabular data and images, simpler classification methods were chosen to be incorporated into the pipeline, and the type of explanations was confined to counterfactual explanations.  Student thesisinfo:eu-repo/semantics/bachelorThesistexthttp://urn.kb.se/resolve?urn=urn:nbn:se:uu:diva-462552IT ; 21120application/pdfinfo:eu-repo/semantics/openAccess
collection NDLTD
language English
format Others
sources NDLTD
topic Engineering and Technology
Teknik och teknologier
spellingShingle Engineering and Technology
Teknik och teknologier
Singh, Vandita
Explainable AI Metrics and Properties for Evaluation and Analysis of Counterfactual Explanations : Explainable AI Metrics and Properties for Evaluation and Analysis of Counterfactual Explanations
description With the recent surge of state-of-the-art AI systems, there has been a growing need to provide explanations for these systems to ensure transparency and understandability by the stakeholders. This has directed the development of increasingly popular Explainable AI systems applicable to a wide range of applications. This trend designates the need for the ability to quantitatively assess and analyze the theoretical and behavioral characteristics of these Explainable AI systems. This study aimed at the following (i) to identify metrics and properties applicable to post-hoc counterfactual explanation methods (a mechanism for generating explanations), (ii) assess the applicability of the identified metrics and properties to compare counterfactual examples, and (iii) analyze the functional and operational characteristics of the selected counterfactual explanation methods. A proof-of-concept tool was designed to perform the given task. A pipeline was implemented, which comprised of selecting a data set, training some suitable classifier, deploying counterfactual generation method, and performing analysis using the metrics selected for implementation. Four counterfactual explanation methods were selected and analyzed against the metrics - distance measures, loss measures, change score values, recourse values in terms of the value of features, and neighborhood. It was found that the metrics provided a means to analyze the characteristics of the selected counterfactual explanation methods. It could also be reasoned which explainers generated better counterfactual explanations for a particular instance by comparing the metrics across different explainers for the same instance. Also, insight could be gained to understand whether a particular counterfactual instance was useful enough or not by comparing the implemented metrics concerning an instance of the desired class. A fact sheet enlisting the functional and operational characteristics of the selected counterfactual methods has been presented. The scope of the work was limited to data sets being comprised of tabular data and images, simpler classification methods were chosen to be incorporated into the pipeline, and the type of explanations was confined to counterfactual explanations. 
author Singh, Vandita
author_facet Singh, Vandita
author_sort Singh, Vandita
title Explainable AI Metrics and Properties for Evaluation and Analysis of Counterfactual Explanations : Explainable AI Metrics and Properties for Evaluation and Analysis of Counterfactual Explanations
title_short Explainable AI Metrics and Properties for Evaluation and Analysis of Counterfactual Explanations : Explainable AI Metrics and Properties for Evaluation and Analysis of Counterfactual Explanations
title_full Explainable AI Metrics and Properties for Evaluation and Analysis of Counterfactual Explanations : Explainable AI Metrics and Properties for Evaluation and Analysis of Counterfactual Explanations
title_fullStr Explainable AI Metrics and Properties for Evaluation and Analysis of Counterfactual Explanations : Explainable AI Metrics and Properties for Evaluation and Analysis of Counterfactual Explanations
title_full_unstemmed Explainable AI Metrics and Properties for Evaluation and Analysis of Counterfactual Explanations : Explainable AI Metrics and Properties for Evaluation and Analysis of Counterfactual Explanations
title_sort explainable ai metrics and properties for evaluation and analysis of counterfactual explanations : explainable ai metrics and properties for evaluation and analysis of counterfactual explanations
publisher Uppsala universitet, Institutionen för informationsteknologi
publishDate 2021
url http://urn.kb.se/resolve?urn=urn:nbn:se:uu:diva-462552
work_keys_str_mv AT singhvandita explainableaimetricsandpropertiesforevaluationandanalysisofcounterfactualexplanationsexplainableaimetricsandpropertiesforevaluationandanalysisofcounterfactualexplanations
_version_ 1723965802311319552