Class incremental learning (CIL) is a challenging setting of continual learning, which learns a series of tasks sequentially. Each task consists of a set of unique classes. The key feature of CIL is that no task identifier (or task-id) is provided at test time. Predicting the task-id for each test sample is a challenging problem. An emerging theory-guided approach (called TIL+OOD) is to train a task-specific model for each task in a shared network for all tasks based on a task-incremental learning (TIL) method to deal with catastrophic forgetting. The model for each task is an out-of-distribution (OOD) detector rather than a conventional classifier. The OOD detector can perform both within-task (in-distribution (IND)) class prediction and OOD detection. The OOD detection capability is the key to task-id prediction during inference. However, this paper argues that using a traditional OOD detector for task-id prediction is sub-optimal because additional information (e.g., the replay data and the learned tasks) available in CIL can be exploited to design a better and principled method for task-id prediction. We call the new method TPL (Task-id Prediction based on Likelihood Ratio). TPL markedly outperforms strong CIL baselines and has negligible catastrophic forgetting. The code of TPL is publicly available at https://github.com/linhaowei1/TPL.
翻译:类增量学习(CIL)是持续学习中一个具有挑战性的设定,它按顺序学习一系列任务。每个任务包含一组独特的类别。CIL的关键特征在于测试时不提供任务标识符(即任务ID)。为每个测试样本预测任务ID是一个难题。一种新兴的理论引导方法(称为TIL+OOD)基于任务增量学习(TIL)方法,在共享网络中为每个任务训练一个特定于任务的模型,以应对灾难性遗忘问题。每个任务的模型是一个离群分布(OOD)检测器,而非传统分类器。该OOD检测器既能进行任务内(分布内(IND))类别预测,也能进行OOD检测。OOD检测能力是推理期间预测任务ID的关键。然而,本文指出,使用传统OOD检测器进行任务ID预测并非最优,因为可以充分利用CIL中可用的额外信息(例如重放数据和已学习任务),设计一种更好且原理性更强的任务ID预测方法。我们称新方法为TPL(基于似然比的任务ID预测)。TPL显著优于强CIL基线方法,且几乎不存在灾难性遗忘。TPL的代码已公开于https://github.com/linhaowei1/TPL。