Post-hoc eXplainable Artificial Intelligence Methods: Counterfactuals and XCBR Applications
Abstract
This thesis addresses the growing need for transparency and user-centric approaches in Explainable Artificial Intelligence (XAI), particularly in applications where AI makes high-stakes decisions. As the adoption of AI accelerates in critical fields, such as healthcare and finance, the need for explanations that are both reliable and understandable has amplified. Counterfactual explanations, which explore «what-if» scenarios, provide actionable insights that align well with human reasoning, making them highly effective for fostering trust and demonstrating causality in AI systems.
This thesis is structured around three core research questions: (1) how integrating domain knowledge and human-centered design can enhance the relevance and trustworthiness of explanations, (2) the development of methods for generating stable, contextually meaningful counterfactuals, and (3) improving comprehensibility and user satisfaction in XAI systems. To address the challenges in these research questions, the thesis introduces several key contributions, including but not limited to PertCF, a perturbation-based method designed for generating counterfactual explanations that balance stability with minimal changes, and the CEval toolkit, a comprehensive evaluation framework developed to assess the fidelity, proximity, diversity, and robustness of instance-based explanations across various metrics.
Empirical studies further validate these contributions, showing that counterfactuals can be tailored to a variety of data types, including tabular and image data while aligning with user preferences when explanations are adaptable to user needs. These findings underscore the value of user-centered XAI, establishing a foundation for future advancements in creating scalable, multimodal, and user-focused explainability methods. The approaches developed in this thesis offer insights for researchers and practitioners working toward transparent and trustworthy AI systems.
Has parts
Paper 1: Bayrak, Betül; Bach, Kerstin. When to Explain? Model Agnostic Explanation Using a Case-based Approach and Counterfactuals. I: Proceedings of the 34th Norwegian ICT conference for research and education – NIKT 2022. OpenProceedings 2023 ISBN 978-3-16-148410-0. s. -.Paper 2: Bayrak, Betül; Bach, Kerstin. A Twin XCBR System Using Supportive and Contrastive Explanations. In proceedings of the ICCBR XCBR’23: Workshop on Case-Based Reasoning for the Explanation of Intelligent Systems at ICCBR2023, Aberdeen, Scotland, 17–20 July 2023.
Paper 3: Bayrak, Betül; Bach, Kerstin. PertCF: A Perturbation-Based Counterfactual Generation Approach. In proceedings of the 43rd SGAI International Conference on Artificial Intelligence, AI, Springer, Cambridge, England. 12–14 December 2023, pp. 174-187.
Paper 4: Bayrak, Betül; Bach, Kerstin. Evaluation of Instance-Based Explanations: An In-Depth Analysis of Counterfactual Evaluation Metrics, Challenges, and the CEval Toolkit. IEEE Access 2024 ;Volum 12. s. Available at: http://dx.doi.org/10.1109/ACCESS.2024.3410540 This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 License.
Paper 5: Darias, Jesus M.; Bayrak, Betül; Caro-Martínez, Marta; Belén, Díaz-Agudo; Recio-Garcia, Juan A.. An Empirical Analysis of User Preferences Regarding XAI Metrics. I: Case-Based Reasoning Research and Development. Springer Berlin/Heidelberg 2024 ISBN 978-3-031-63645-5. s. 96-110. Available at: http://dx.doi.org/https://doi.org/10.1007/978-3-031-63646-2_7