Explainable Artificial Intelligence (XAI) seeks to elucidate the decision-making mechanisms of AI models, enabling users to glean insights beyond the results they produce. While a key objective of XAI is to enhance the performance of AI models through explanatory processes, a notable portion of XAI literature predominantly addresses the explanation of AI systems, with limited focus on leveraging XAI methods for performance improvement. This study introduces a novel approach utilizing Integrated Gradients explanations to enhance a classification system, which is subsequently evaluated on three datasets: Fashion-MNIST, CIFAR10, and STL10. Empirical findings indicate that Integrated Gradients explanations effectively contribute to enhancing classification performance.

An XAI-based masking approach to improve classification systems / Apicella, A.; Giugliano, S.; Isgro, F.; Pollastro, A.; Prevete, R.. - 3615:(2023), pp. 79-83. ( 2nd Workshop on Bias, Ethical AI, Explainability and the Role of Logic and Logic Programming, BEWARE 2023 ita 2023).

An XAI-based masking approach to improve classification systems

Apicella A.;Isgro F.;Prevete R.
2023

Abstract

Explainable Artificial Intelligence (XAI) seeks to elucidate the decision-making mechanisms of AI models, enabling users to glean insights beyond the results they produce. While a key objective of XAI is to enhance the performance of AI models through explanatory processes, a notable portion of XAI literature predominantly addresses the explanation of AI systems, with limited focus on leveraging XAI methods for performance improvement. This study introduces a novel approach utilizing Integrated Gradients explanations to enhance a classification system, which is subsequently evaluated on three datasets: Fashion-MNIST, CIFAR10, and STL10. Empirical findings indicate that Integrated Gradients explanations effectively contribute to enhancing classification performance.
2023
An XAI-based masking approach to improve classification systems / Apicella, A.; Giugliano, S.; Isgro, F.; Pollastro, A.; Prevete, R.. - 3615:(2023), pp. 79-83. ( 2nd Workshop on Bias, Ethical AI, Explainability and the Role of Logic and Logic Programming, BEWARE 2023 ita 2023).
File in questo prodotto:
File Dimensione Formato  
short2.pdf

accesso aperto

Licenza: Creative commons
Dimensione 1.06 MB
Formato Adobe PDF
1.06 MB Adobe PDF Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11588/963552
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 2
  • ???jsp.display-item.citation.isi??? ND
social impact