Diffusion models have led to significant advancements in generative modelling. Yet their widespread adoption poses challenges regarding data attribution and interpretability. In this paper, we aim to help address such challenges in diffusion models by developing an influence functions framework. Influence function-based data attribution methods approximate how a model's output would have changed if some training data were removed. In supervised learning, this is usually used for predicting how the loss on a particular example would change. For diffusion models, we focus on predicting the change in the probability of generating a particular example via several proxy measurements. We show how to formulate influence functions for such quantities and how previously proposed methods can be interpreted as particular design choices in our framework. To ensure scalability of the Hessian computations in influence functions, we systematically develop K-FAC approximations based on generalised Gauss-Newton matrices specifically tailored to diffusion models. We recast previously proposed methods as specific design choices in our framework and show that our recommended method outperforms previous data attribution approaches on common evaluations, such as the Linear Data-modelling Score (LDS) or retraining without top influences, without the need for method-specific hyperparameter tuning.
翻译:扩散模型在生成建模领域取得了显著进展。然而,其广泛采用带来了数据归因与可解释性方面的挑战。本文旨在通过构建影响函数框架,帮助应对扩散模型中的此类挑战。基于影响函数的数据归因方法近似估计移除部分训练数据时模型输出的变化。在监督学习中,该方法通常用于预测特定样本的损失如何变化。对于扩散模型,我们重点通过若干代理度量来预测生成特定样本的概率变化。我们展示了如何为此类量值构建影响函数,并说明先前提出的方法如何可解释为我们框架中的特定设计选择。为确保影响函数中Hessian矩阵计算的可扩展性,我们基于专门为扩散模型定制的广义高斯-牛顿矩阵,系统性地开发了K-FAC近似方法。我们将先前提出的方法重新阐释为我们框架中的特定设计选择,并证明在常见评估指标(如线性数据建模评分LDS或移除高影响力样本的重训练)上,我们推荐的方法优于以往的数据归因方法,且无需针对特定方法进行超参数调优。