We advocate for a new statistical principle that combines the most desirable aspects of both parameter inference and density estimation. This leads us to the predictively oriented (PrO) posterior, which expresses uncertainty as a consequence of predictive ability. Doing so leads to inferences which predictively dominate both classical and generalised Bayes posterior predictive distributions: up to logarithmic factors, PrO posteriors converge to the predictively optimal model average. Whereas classical and generalised Bayes posteriors only achieve this rate if the model can recover the data-generating process, PrO posteriors adapt to the level of model misspecification. This means that they concentrate around the true model in the same way as Bayes and Gibbs posteriors if the model can recover the data-generating distribution, but do not concentrate in the presence of non-trivial forms of model misspecification. Instead, they stabilise towards a predictively optimal posterior whose degree of irreducible uncertainty admits an interpretation as the degree of model misspecification -- a sharp contrast to how Bayesian uncertainty and its existing extensions behave. Lastly, we show that PrO posteriors can be sampled from by evolving particles based on mean field Langevin dynamics, and verify the practical significance of our theoretical developments on a number of numerical examples.
翻译:我们倡导一种新的统计原则,该原则结合了参数推断与密度估计中最理想的特性。这引导我们提出了预测导向(PrO)后验,它将不确定性表达为预测能力的结果。由此得到的推断在预测性能上优于经典贝叶斯和广义贝叶斯后验预测分布:在忽略对数因子的意义下,PrO后验收敛于预测最优的模型平均。经典与广义贝叶斯后验仅当模型能够还原数据生成过程时才能达到此收敛速率,而PrO后验能够自适应模型误设的程度。这意味着,若模型能够还原数据生成分布,PrO后验会以与贝叶斯后验和吉布斯后验相同的方式集中于真实模型;但在存在非平凡模型误设的情况下,它们不会过度集中。相反,它们会稳定趋近于一个预测最优的后验,其不可约不确定性的程度可解释为模型误设的程度——这与贝叶斯不确定性及其现有扩展的行为形成鲜明对比。最后,我们证明可以通过基于平均场朗之万动力学演化粒子来对PrO后验进行采样,并通过若干数值算例验证了我们理论发展的实际意义。