The increasing demand for transparent and reliable models, particularly in high-stakes decision-making areas such as medical image analysis, has led to the emergence of eXplainable Artificial Intelligence (XAI). Post-hoc XAI techniques, which aim to explain black-box models after training, have raised concerns about their fidelity to model predictions. In contrast, Self-eXplainable AI (S-XAI) offers a compelling alternative by incorporating explainability directly into the training process of deep learning models. This approach allows models to generate inherent explanations that are closely aligned with their internal decision-making processes, enhancing transparency and supporting the trustworthiness, robustness, and accountability of AI systems in real-world medical applications. To facilitate the development of S-XAI methods for medical image analysis, this survey presents a comprehensive review across various image modalities and clinical applications. It covers more than 200 papers from three key perspectives: 1) input explainability through the integration of explainable feature engineering and knowledge graph, 2) model explainability via attention-based learning, concept-based learning, and prototype-based learning, and 3) output explainability by providing textual and counterfactual explanations. This paper also outlines desired characteristics of explainability and evaluation methods for assessing explanation quality, while discussing major challenges and future research directions in developing S-XAI for medical image analysis.
翻译:对透明可靠模型日益增长的需求,特别是在医学影像分析等高风险决策领域,推动了可解释人工智能(XAI)的兴起。旨在训练后解释黑盒模型的事后XAI技术,其解释与模型预测的真实性已引发担忧。相比之下,自解释人工智能(S-XAI)提供了一种引人注目的替代方案,它将可解释性直接融入深度学习模型的训练过程。这种方法使模型能够生成与其内部决策过程紧密一致的内在解释,从而增强透明度,并支持AI系统在实际医疗应用中的可信性、鲁棒性和可问责性。为促进医学影像分析中S-XAI方法的发展,本综述对不同影像模态和临床应用进行了全面回顾。它从三个关键视角涵盖了200多篇论文:1)通过可解释特征工程与知识图谱融合实现输入可解释性;2)通过基于注意力的学习、基于概念的学习和基于原型的学习实现模型可解释性;3)通过提供文本解释和反事实解释实现输出可解释性。本文还概述了可解释性的理想特性与评估解释质量的评价方法,同时探讨了开发用于医学影像分析的S-XAI所面临的主要挑战与未来研究方向。