The equivalence of realizable and agnostic learnability is a fundamental phenomenon in learning theory. With variants ranging from classical settings like PAC learning and regression to recent trends such as adversarially robust learning, it's surprising that we still lack a unified theory; traditional proofs of the equivalence tend to be disparate, and rely on strong model-specific assumptions like uniform convergence and sample compression. In this work, we give the first model-independent framework explaining the equivalence of realizable and agnostic learnability: a three-line blackbox reduction that simplifies, unifies, and extends our understanding across a wide variety of settings. This includes models with no known characterization of learnability such as learning with arbitrary distributional assumptions and more general loss functions, as well as a host of other popular settings such as robust learning, partial learning, fair learning, and the statistical query model. More generally, we argue that the equivalence of realizable and agnostic learning is actually a special case of a broader phenomenon we call property generalization: any desirable property of a learning algorithm (e.g. noise tolerance, privacy, stability) that can be satisfied over finite hypothesis classes extends (possibly in some variation) to any learnable hypothesis class.
翻译:可实现学习与不可知学习的等价性是学习理论中的一个基本现象。从经典的PAC学习、回归等传统设定,到近期如对抗鲁棒学习等新趋势,令人惊讶的是我们仍缺乏统一理论;传统的等价性证明往往相互孤立,且依赖于强模型特定假设(如一致收敛和样本压缩)。在这项工作中,我们首次提出了解释可实现与不可知学习等价性的模型无关框架:一个三行黑盒归约,它简化、统一并扩展了我们在广泛设定中的理解。这包括尚未有学习性表征的模型(如任意分布假设下的学习与更通用的损失函数),以及其他众多流行设定(如鲁棒学习、部分学习、公平学习和统计查询模型)。更一般地,我们认为可实现学习与不可知学习的等价性实际上是更广泛现象(我们称之为属性泛化)的特例:任何学习算法的理想属性(例如噪声容忍性、隐私性、稳定性),若能在有限假设类上满足,则可(可能以某种变体形式)推广到任何可学习的假设类别。