Generalized Additive Models (GAMs) are commonly considered *interpretable* within the ML community, as their structure makes the relationship between inputs and outputs relatively understandable. Therefore, it may seem natural to hypothesize that obtaining meaningful explanations for GAMs could be performed efficiently and would not be computationally infeasible. In this work, we challenge this hypothesis by analyzing the *computational complexity* of generating different explanations for various forms of GAMs across multiple contexts. Our analysis reveals a surprisingly diverse landscape of both positive and negative complexity outcomes. Particularly, under standard complexity assumptions such as P!=NP, we establish several key findings: (1) in stark contrast to many other common ML models, the complexity of generating explanations for GAMs is heavily influenced by the structure of the input space; (2) the complexity of explaining GAMs varies significantly with the types of component models used - but interestingly, these differences only emerge under specific input domain settings; (3) significant complexity distinctions appear for obtaining explanations in regression tasks versus classification tasks in GAMs; and (4) expressing complex models like neural networks additively (e.g., as neural additive models) can make them easier to explain, though interestingly, this benefit appears only for certain explanation methods and input domains. Collectively, these results shed light on the feasibility of computing diverse explanations for GAMs, offering a rigorous theoretical picture of the conditions under which such computations are possible or provably hard.
翻译:广义可加性模型(GAMs)在机器学习领域通常被视为*可解释*模型,因其结构使得输入与输出之间的关系相对易于理解。因此,一个看似自然的假设是:为GAMs获取有意义的解释可能高效可行,且不会在计算上不可实现。本研究通过分析在不同情境下为各类GAMs生成不同解释的*计算复杂性*,对这一假设提出了挑战。我们的分析揭示了正负复杂性结果并存的惊人多样化图景。特别是在P≠NP等标准复杂性假设下,我们确立了若干关键发现:(1)与许多其他常见机器学习模型形成鲜明对比的是,为GAMs生成解释的复杂性受输入空间结构的显著影响;(2)解释GAMs的复杂性随所用组件模型类型的不同而显著变化——但有趣的是,这些差异仅在特定输入域设置下才会显现;(3)在回归任务与分类任务中为GAMs获取解释存在显著的复杂性差异;(4)将神经网络等复杂模型以可加形式表达(例如神经可加性模型)可使其更易于解释,但有趣的是,这种优势仅在某些解释方法和输入域中显现。总体而言,这些结果阐明了为GAMs计算多样化解释的可行性,为理解此类计算在何种条件下可能实现或可证明困难提供了严谨的理论框架。