论文标题
Imagenet的大规模开放式分类协议
Large-Scale Open-Set Classification Protocols for ImageNet
论文作者
论文摘要
开放式分类(OSC)打算将封闭式分类模型调整到现实世界中的情况下,其中分类器必须正确标记已知类别的样本,同时拒绝以前未见的未知样本。直到最近,研究才开始研究能够正确处理这些未知样本的算法。其中一些方法通过将分类器学会拒绝的训练集负面样本包括在训练集中,以期待这些数据会增加分类器在未知类别上的鲁棒性。这些方法中的大多数都在小型和低分辨率图像数据集(如MNIST,SVHN或CIFAR)上进行评估,这使得很难评估其对现实世界的适用性,并彼此比较。我们提出了三个开放式协议,这些协议为已知类别和未知类别之间具有不同级别的相似性提供丰富的自然图像数据集。协议由选定的ImageNet类的子集组成,以提供培训和测试数据更接近现实世界情景。此外,我们提出了一个新的验证指标,可以用来评估深度学习模型的培训是否解决已知样本的分类和拒绝未知样本。我们使用协议将两种基线开放式算法的性能与标准软马克斯基线进行比较,并发现该算法在训练过程中已经看到的负样本效果很好,并且部分在脱离分布检测任务上,但是在以前未知类别的样品存在下降低了性能。
Open-Set Classification (OSC) intends to adapt closed-set classification models to real-world scenarios, where the classifier must correctly label samples of known classes while rejecting previously unseen unknown samples. Only recently, research started to investigate on algorithms that are able to handle these unknown samples correctly. Some of these approaches address OSC by including into the training set negative samples that a classifier learns to reject, expecting that these data increase the robustness of the classifier on unknown classes. Most of these approaches are evaluated on small-scale and low-resolution image datasets like MNIST, SVHN or CIFAR, which makes it difficult to assess their applicability to the real world, and to compare them among each other. We propose three open-set protocols that provide rich datasets of natural images with different levels of similarity between known and unknown classes. The protocols consist of subsets of ImageNet classes selected to provide training and testing data closer to real-world scenarios. Additionally, we propose a new validation metric that can be employed to assess whether the training of deep learning models addresses both the classification of known samples and the rejection of unknown samples. We use the protocols to compare the performance of two baseline open-set algorithms to the standard SoftMax baseline and find that the algorithms work well on negative samples that have been seen during training, and partially on out-of-distribution detection tasks, but drop performance in the presence of samples from previously unseen unknown classes.