The denoising diffusion probabilistic model (DDPM) has emerged as a mainstream generative model in generative AI. While sharp convergence guarantees have been established for the DDPM, the iteration complexity is, in general, proportional to the ambient data dimension, resulting in overly conservative theory that fails to explain its practical efficiency. This has motivated the recent work Li and Yan (2024a) to investigate how the DDPM can achieve sampling speed-ups through automatic exploitation of intrinsic low dimensionality of data. We strengthen this line of work by demonstrating, in some sense, optimal adaptivity to unknown low dimensionality. For a broad class of data distributions with intrinsic dimension $k$, we prove that the iteration complexity of the DDPM scales nearly linearly with $k$, which is optimal when using KL divergence to measure distributional discrepancy. Notably, our work is closely aligned with the independent concurrent work Potaptchik et al. (2024) -- posted two weeks prior to ours -- in establishing nearly linear-$k$ convergence guarantees for the DDPM.
翻译:去噪扩散概率模型(DDPM)已成为生成式人工智能中的主流生成模型。虽然DDPM已建立了严格的收敛性保证,但其迭代复杂度通常与数据的环境维度成正比,导致理论过于保守,无法解释其实际效率。这促使近期工作Li and Yan (2024a) 研究DDPM如何通过自动利用数据的内在低维性来实现采样加速。我们通过证明其在某种意义上对未知低维度的最优适应性,加强了这一研究方向。对于一类具有内在维度$k$的广泛数据分布,我们证明了DDPM的迭代复杂度与$k$近似呈线性关系,这在用KL散度衡量分布差异时是最优的。值得注意的是,我们的工作与独立同期工作Potaptchik et al. (2024)——比我们的工作早发布两周——在建立DDPM近似线性$k$收敛保证方面高度一致。