Score-based generative models have recently achieved remarkable success. While they are usually parameterized by the score, an alternative way is to use a series of time-dependent energy-based models (EBMs), where the score is obtained from the negative input-gradient of the energy. Crucially, EBMs can be leveraged not only for generation, but also for tasks such as compositional sampling or building Boltzmann Generators via Monte Carlo methods. However, training EBMs remains challenging. Direct maximum likelihood is computationally prohibitive due to the need for nested sampling, while score matching, though efficient, suffers from mode blindness. To address these issues, we introduce the Diffusive Classification (DiffCLF) objective, a simple method that avoids blindness while remaining computationally efficient. DiffCLF reframes EBM learning as a supervised classification problem across noise levels, and can be seamlessly combined with standard score-based objectives. We validate the effectiveness of DiffCLF by comparing the estimated energies against ground truth in analytical Gaussian mixture cases, and by applying the trained models to tasks such as model composition and Boltzmann Generator sampling. Our results show that DiffCLF enables EBMs with higher fidelity and broader applicability than existing approaches.
翻译:基于分数的生成模型近期取得了显著成功。虽然它们通常通过分数进行参数化,但另一种方法是使用一系列时间依赖的基于能量的模型(EBMs),其中分数通过能量的负输入梯度获得。关键在于,EBMs不仅可用于生成,还可用于组合采样或通过蒙特卡洛方法构建玻尔兹曼生成器等任务。然而,训练EBMs仍然具有挑战性。由于需要嵌套采样,直接最大似然估计在计算上不可行;而分数匹配虽然高效,却存在模式盲区问题。为解决这些问题,我们引入了扩散分类(DiffCLF)目标,这是一种避免盲区同时保持计算效率的简单方法。DiffCLF将EBM学习重新定义为跨噪声水平的监督分类问题,并可无缝结合标准的基于分数的目标。我们通过在解析高斯混合案例中将估计能量与真实值进行比较,并将训练模型应用于模型组合和玻尔兹曼生成器采样等任务,验证了DiffCLF的有效性。我们的结果表明,与现有方法相比,DiffCLF能够实现更高保真度和更广泛适用性的EBMs。