We consider estimators obtained by iterates of the conjugate gradient (CG) algorithm applied to the normal equation of prototypical statistical inverse problems. Stopping the CG algorithm early induces regularisation, and optimal convergence rates of prediction and reconstruction error are established in wide generality for an ideal oracle stopping time. Based on this insight, a fully data-driven early stopping rule $\tau$ is constructed, which also attains optimal rates, provided the error in estimating the noise level is not dominant. The error analysis of CG under statistical noise is subtle due to its nonlinear dependence on the observations. We provide an explicit error decomposition into two terms, which shares important properties of the classical bias-variance decomposition. Together with a continuous interpolation between CG iterates, this paves the way for a comprehensive error analysis of early stopping. In particular, a general oracle-type inequality is proved for the prediction error at $\tau$. For bounding the reconstruction error, a more refined probabilistic analysis, based on concentration of self-normalised Gaussian processes, is developed. The methodology also provides some new insights into early stopping for CG in deterministic inverse problems. A numerical study for standard examples shows good results in practice for early stopping at $\tau$.
翻译:我们考虑将共轭梯度(CG)算法应用于典型统计逆问题正规方程迭代所获得的估计量。提前停止CG算法会引入正则化效应,我们为理想的神谕停止时间建立了预测误差与重构误差的最优收敛速率普适理论。基于这一认识,我们构建了一种完全数据驱动的提前停止规则$\tau$,该规则在噪声水平估计误差不占主导地位时同样能达到最优速率。由于CG算法对观测数据的非线性依赖特性,其在统计噪声下的误差分析较为复杂。我们给出了误差分解为两项的显式表达式,该分解具有经典偏差-方差分解的重要特性。结合CG迭代间的连续插值技术,这为提前停止的全面误差分析奠定了基础。特别地,我们证明了$\tau$处预测误差的广义神谕型不等式。为界定重构误差,我们基于自归一化高斯过程的集中性理论,发展了更精细的概率分析方法。该方法也为确定性逆问题中CG的提前停止提供了新的理论视角。对标准算例的数值研究表明,在$\tau$处实施提前停止在实践中能获得良好效果。