论文标题
Rankme:评估预定的自我监督陈述的下游表现
RankMe: Assessing the downstream performance of pretrained self-supervised representations by their rank
论文作者
论文摘要
随着许多方法变化的出现,联合安装自我监督学习(JE-SSL)的发展迅速发展,但只有很少有原则指南可以帮助从业者成功部署它们。陷阱的主要原因是JE-SSL的核心原则,即不采用任何输入重建,因此缺乏失败的培训的视觉提示。添加非信息性损失值,很难将SSL部署在一个新数据集中,没有标签可以帮助判断学习的表示的质量。在这项研究中,我们开发了一个简单的无监督标准,该标准指示了学识渊博的JE-SSL表示的质量:它们的有效等级。尽管简单且计算友好,但这种方法(即置换的rankme)允许人们评估JE-SSL表示的性能,即使在不同的下游数据集中,也可以评估无需任何标签的情况。 Rankme的另一个好处是,它没有任何培训或超参数可以调节。通过涉及数百个训练发作的彻底的经验实验,我们演示了如何将RankMe用于超参数选择,而与涉及数据集标签的当前选择方法相比,最终性能几乎没有降低。我们希望RankMe将促进JE-SSL的部署到没有机会依靠标签来代表质量评估的领域。
Joint-Embedding Self Supervised Learning (JE-SSL) has seen a rapid development, with the emergence of many method variations but only few principled guidelines that would help practitioners to successfully deploy them. The main reason for that pitfall comes from JE-SSL's core principle of not employing any input reconstruction therefore lacking visual cues of unsuccessful training. Adding non informative loss values to that, it becomes difficult to deploy SSL on a new dataset for which no labels can help to judge the quality of the learned representation. In this study, we develop a simple unsupervised criterion that is indicative of the quality of the learned JE-SSL representations: their effective rank. Albeit simple and computationally friendly, this method -- coined RankMe -- allows one to assess the performance of JE-SSL representations, even on different downstream datasets, without requiring any labels. A further benefit of RankMe is that it does not have any training or hyper-parameters to tune. Through thorough empirical experiments involving hundreds of training episodes, we demonstrate how RankMe can be used for hyperparameter selection with nearly no reduction in final performance compared to the current selection method that involve a dataset's labels. We hope that RankMe will facilitate the deployment of JE-SSL towards domains that do not have the opportunity to rely on labels for representations' quality assessment.