论文标题
一致性是进一步减轻持续学习中灾难性遗忘的关键
Consistency is the key to further mitigating catastrophic forgetting in continual learning
论文作者
论文摘要
深层神经网络由于灾难性忘记了以前学到的任务而难以继续学习多个顺序任务。基于排练的方法将以前的任务样本明确存储在缓冲区中,并将其与当前的任务样本交织在一起,这被证明是缓解遗忘的最有效的方法。但是,由于其性能与缓冲区的大小相称,因此在低缓冲机制和更长的任务序列下,体验重播(ER)表现不佳。软目标预测的一致性可以帮助ER保存与先前任务有关的信息,因为软目标捕获了数据的丰富相似性结构。因此,我们研究了在各种持续学习方案下,一致性正则化在ER框架中的作用。我们还建议将一致性正规化作为一个自制的借口任务,从而使使用各种自我监督的学习方法作为正规化者。尽管同时增强了对自然腐败的模型校准和鲁棒性,但在预测中正规化一致性会导致在所有持续学习场景中遗忘。在不同的正规化家族中,我们发现更严格的一致性约束将以前的任务信息更好地保留。
Deep neural networks struggle to continually learn multiple sequential tasks due to catastrophic forgetting of previously learned tasks. Rehearsal-based methods which explicitly store previous task samples in the buffer and interleave them with the current task samples have proven to be the most effective in mitigating forgetting. However, Experience Replay (ER) does not perform well under low-buffer regimes and longer task sequences as its performance is commensurate with the buffer size. Consistency in predictions of soft-targets can assist ER in preserving information pertaining to previous tasks better as soft-targets capture the rich similarity structure of the data. Therefore, we examine the role of consistency regularization in ER framework under various continual learning scenarios. We also propose to cast consistency regularization as a self-supervised pretext task thereby enabling the use of a wide variety of self-supervised learning methods as regularizers. While simultaneously enhancing model calibration and robustness to natural corruptions, regularizing consistency in predictions results in lesser forgetting across all continual learning scenarios. Among the different families of regularizers, we find that stricter consistency constraints preserve previous task information in ER better.