Dealing with distribution shifts is one of the central challenges for modern machine learning. One fundamental situation is the covariate shift, where the input distributions of data change from training to testing stages while the input-conditional output distribution remains unchanged. In this paper, we initiate the study of a more challenging scenario -- continuous covariate shift -- in which the test data appear sequentially, and their distributions can shift continuously. Our goal is to adaptively train the predictor such that its prediction risk accumulated over time can be minimized. Starting with the importance-weighted learning, we show the method works effectively if the time-varying density ratios of test and train inputs can be accurately estimated. However, existing density ratio estimation methods would fail due to data scarcity at each time step. To this end, we propose an online method that can appropriately reuse historical information. Our density ratio estimation method is proven to perform well by enjoying a dynamic regret bound, which finally leads to an excess risk guarantee for the predictor. Empirical results also validate the effectiveness.
翻译:处理分布偏移是现代机器学习的核心挑战之一。协变量偏移是一种基本情形,即数据的输入分布在训练阶段与测试阶段之间发生变化,而输入条件输出分布保持不变。本文首次研究了一个更具挑战性的场景——连续协变量偏移——其中测试数据按序出现,且其分布可能连续变化。我们的目标是自适应地训练预测器,以最小化其随时间累积的预测风险。从重要性加权学习出发,我们证明若能够准确估计测试与训练输入的时变密度比,该方法将有效工作。然而,由于每个时间步的数据稀缺,现有的密度比估计方法将失效。为此,我们提出一种能够合理复用历史信息的在线方法。我们的密度比估计方法被证明具有动态遗憾界,从而最终为预测器提供了超额风险保证。实证结果也验证了其有效性。