In the field of medical imaging, particularly in tasks related to early disease detection and prognosis, understanding the reasoning behind AI model predictions is imperative for assessing their reliability. Conventional explanation methods encounter challenges in identifying decisive features in medical image classifications, especially when discriminative features are subtle or not immediately evident. To address this limitation, we propose an agent model capable of generating counterfactual images that prompt different decisions when plugged into a black box model. By employing this agent model, we can uncover influential image patterns that impact the black model's final predictions. Through our methodology, we efficiently identify features that influence decisions of the deep black box. We validated our approach in the rigorous domain of medical prognosis tasks, showcasing its efficacy and potential to enhance the reliability of deep learning models in medical image classification compared to existing interpretation methods. The code will be publicly available at https://github.com/ayanglab/DiffExplainer.
翻译:在医学影像领域,特别是与早期疾病检测和预后相关的任务中,理解AI模型预测背后的推理对于评估其可靠性至关重要。传统的解释方法在识别医学图像分类中的决定性特征时面临挑战,尤其是当判别性特征较为细微或不明显时。为应对这一局限,我们提出了一种代理模型,该模型能够生成反事实图像,当输入黑盒模型时会导致不同的决策。通过使用该代理模型,我们可以揭示影响黑盒模型最终预测的关键图像模式。通过我们的方法,我们高效地识别了影响深度黑盒模型决策的特征。我们在严格的医学预后任务领域中验证了我们的方法,展示了其与现有解释方法相比,在提升医学图像分类深度学习模型可靠性方面的有效性和潜力。代码将在 https://github.com/ayanglab/DiffExplainer 公开提供。