Artificial Intelligence decision-making systems have dramatically increased their predictive performance in recent years, beating humans in many different specific tasks. However, with increased performance has come an increase in the complexity of the black-box models adopted by the AI systems, making them entirely obscure for the decision process adopted. Explainable AI is a field that seeks to make AI decisions more transparent by producing explanations. In this paper, we propose T-LACE, an approach able to retrieve post-hoc counterfactual explanations for a given pre-trained black-box model. T-LACE exploits the similarity and linearity proprieties of a custom-created transparent latent space to build reliable counterfactual explanations. We tested T-LACE on several tabular datasets and provided qualitative evaluations of the generated explanations in terms of similarity, robustness, and diversity. Comparative analysis against various state-of-the-art counterfactual explanation methods shows the higher effectiveness of our approach.

Transparent Latent Space Counterfactual Explanations for Tabular Data

Bodria, Francesco;Giannotti, Fosca;Pedreschi, Dino
2022

Abstract

Artificial Intelligence decision-making systems have dramatically increased their predictive performance in recent years, beating humans in many different specific tasks. However, with increased performance has come an increase in the complexity of the black-box models adopted by the AI systems, making them entirely obscure for the decision process adopted. Explainable AI is a field that seeks to make AI decisions more transparent by producing explanations. In this paper, we propose T-LACE, an approach able to retrieve post-hoc counterfactual explanations for a given pre-trained black-box model. T-LACE exploits the similarity and linearity proprieties of a custom-created transparent latent space to build reliable counterfactual explanations. We tested T-LACE on several tabular datasets and provided qualitative evaluations of the generated explanations in terms of similarity, robustness, and diversity. Comparative analysis against various state-of-the-art counterfactual explanation methods shows the higher effectiveness of our approach.
2022
Settore INF/01 - Informatica
9th IEEE International Conference on Data Science and Advanced Analytics, DSAA 2022
China
2022
Proceedings - 2022 IEEE 9th International Conference on Data Science and Advanced Analytics, DSAA 2022
Institute of Electrical and Electronics Engineers Inc.
978-1-6654-7330-9
Counterfactuals Explanation; Explainable Artificial Intelligence; Latent Space
   SoBigData++: European Integrated Infrastructure for Social Mining and Big Data Analytics
   SoBigData-PlusPlus
   European Commission
   Horizon 2020 Framework Programme
   871042

   HumanE AI Network
   HumanE-AI-Net
   European Commission
   Horizon 2020 Framework Programme
   952026

   Science and technology for the explanation of AI decision making
   XAI
   European Commission
   Horizon 2020 Framework Programme
   834756

   Foundations of Trustworthy AI - Integrating Reasoning, Learning and Optimization
   TAILOR
   European Commission
   Horizon 2020 Framework Programme
   952215
ACM SIGKDD
File in questo prodotto:
File Dimensione Formato  
Transparent_Latent_Space_Counterfactual_Explanations_for_Tabular_Data.pdf

Accesso chiuso

Tipologia: Published version
Licenza: Non pubblico
Dimensione 3.41 MB
Formato Adobe PDF
3.41 MB Adobe PDF   Richiedi una copia

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11384/137134
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 1
  • ???jsp.display-item.citation.isi??? 0
social impact