While Explainable AI (XAI) helps users understand AI decisions, misalignment in domain knowledge can lead to disagreement. This inconsistency hinders understanding, and because explanations are often read-only, users lack the control to improve alignment. We propose making XAI editable, allowing users to write rules to improve control and gain deeper understanding through the generation effect of active learning. We developed CoExplain, leveraging a neural network for universal representation and symbolic rules for intuitive reasoning on interpretable attributes. CoExplain explains the neural network with a faithful proxy decision tree, parses user-written rules as an equivalent neural network graph, and collaboratively optimizes the decision tree. In a user study (N=43), CoExplain and manually editable XAI improved user understanding and model alignment compared to read-only XAI. CoExplain was easier to use with fewer edits and less time. This work contributes Editable XAI for bidirectional AI alignment, improving understanding and control.
翻译:尽管可解释人工智能(XAI)能帮助用户理解AI决策,但领域知识的不对齐可能导致认知分歧。这种不一致性会阻碍理解过程,且由于解释通常为只读形式,用户缺乏改善对齐的控制能力。我们提出使XAI具备可编辑性,允许用户通过编写规则来增强控制,并借助主动学习的生成效应获得更深层次的理解。我们开发了CoExplain系统,该系统利用神经网络实现通用表征,并采用符号规则对可解释属性进行直观推理。CoExplain通过忠实代理决策树解释神经网络,将用户编写的规则解析为等效的神经网络计算图,并协同优化决策树结构。在用户研究(N=43)中,相较于只读式XAI,CoExplain与手动可编辑XAI均显著提升了用户理解度与模型对齐度。CoExplain以更少的编辑次数和更短的时间成本展现出更优的易用性。本研究提出的可编辑XAI为实现双向AI对齐提供了新范式,有效提升了理解效能与控制能力。