论文标题
一项对不平衡类增量学习校准方法的比较研究
A Comparative Study of Calibration Methods for Imbalanced Class Incremental Learning
论文作者
论文摘要
深度学习方法在广泛的AI问题中是成功的,尤其是在视觉识别任务中。但是,其中仍然存在开放的问题,其中有能力处理视觉信息流以及数据集中的类不平衡的管理。现有研究在现实世界应用中共同发生时分别对这两个问题进行了处理。在这里,我们从不平衡的数据集中逐步研究学习问题。我们专注于具有恒定的深层模型复杂性的算法,并使用有界的内存来存储跨增量状态的旧类的示例。由于内存是有限的,因此学习旧类的图像少于新类,并且由于增量学习而导致的不平衡,将添加到初始数据集不平衡中。出现了有利于新类别的分数预测偏见,我们评估了一组综合的得分校准方法来减少它。评估是使用三个数据集进行的,使用两个数据集不平衡配置和三个有界内存大小。结果表明,大多数校准方法具有有益的效果,并且对于下限内存量最有用,这在实践中最有趣。作为次要贡献,我们从增量学习算法的损失函数中删除了通常的蒸馏成分。我们表明,简单的香草微调是一种不平衡的递增学习算法的骨干。
Deep learning approaches are successful in a wide range of AI problems and in particular for visual recognition tasks. However, there are still open problems among which is the capacity to handle streams of visual information and the management of class imbalance in datasets. Existing research approaches these two problems separately while they co-occur in real world applications. Here, we study the problem of learning incrementally from imbalanced datasets. We focus on algorithms which have a constant deep model complexity and use a bounded memory to store exemplars of old classes across incremental states. Since memory is bounded, old classes are learned with fewer images than new classes and an imbalance due to incremental learning is added to the initial dataset imbalance. A score prediction bias in favor of new classes appears and we evaluate a comprehensive set of score calibration methods to reduce it. Evaluation is carried with three datasets, using two dataset imbalance configurations and three bounded memory sizes. Results show that most calibration methods have beneficial effect and that they are most useful for lower bounded memory sizes, which are most interesting in practice. As a secondary contribution, we remove the usual distillation component from the loss function of incremental learning algorithms. We show that simpler vanilla fine tuning is a stronger backbone for imbalanced incremental learning algorithms.