论文标题
通过分组和自我注意来抑制错误标签的数据
Suppressing Mislabeled Data via Grouping and Self-Attention
论文作者
论文摘要
深网络在大规模清洁数据上取得了出色的成果,但是从嘈杂的标签中学习时会大大降低。为了抑制错误标记的数据的影响,本文提出了一个概念上简单但有效的训练块,称为“细心特征混合”(AFM),这使得通过小组的样本相互作用更加注意清洁样品,而较少地将其标记为错误。具体来说,该插件AFM首先利用A \ textit {group-to-Atternend}模块来构建组并为小组样本分配注意力权重,然后使用\ textit {Mixup}模块,并带有注意力重量来插入互构层噪音噪声噪声。 AFM可以为噪音的深度学习带来一些吸引人的好处。 (i)它不依赖任何假设和额外的干净子集。 (ii)通过大量插值,与原始噪声比相比,无用样品的比率大大降低。 (iii)\ pxj {它可以通过分类器共同优化插值权重,从而通过低注意力重量抑制了错误标记的数据的影响。 (iv) It partially inherits the vicinal risk minimization of mixup to alleviate over-fitting while improves it by sampling fewer feature-target vectors around mislabeled data from the mixup vicinal distribution.} Extensive experiments demonstrate that AFM yields state-of-the-art results on two challenging real-world noisy datasets: Food101N and Clothing1M.该代码将在https://github.com/kaiwang960112/afm上找到。
Deep networks achieve excellent results on large-scale clean data but degrade significantly when learning from noisy labels. To suppressing the impact of mislabeled data, this paper proposes a conceptually simple yet efficient training block, termed as Attentive Feature Mixup (AFM), which allows paying more attention to clean samples and less to mislabeled ones via sample interactions in small groups. Specifically, this plug-and-play AFM first leverages a \textit{group-to-attend} module to construct groups and assign attention weights for group-wise samples, and then uses a \textit{mixup} module with the attention weights to interpolate massive noisy-suppressed samples. The AFM has several appealing benefits for noise-robust deep learning. (i) It does not rely on any assumptions and extra clean subset. (ii) With massive interpolations, the ratio of useless samples is reduced dramatically compared to the original noisy ratio. (iii) \pxj{It jointly optimizes the interpolation weights with classifiers, suppressing the influence of mislabeled data via low attention weights. (iv) It partially inherits the vicinal risk minimization of mixup to alleviate over-fitting while improves it by sampling fewer feature-target vectors around mislabeled data from the mixup vicinal distribution.} Extensive experiments demonstrate that AFM yields state-of-the-art results on two challenging real-world noisy datasets: Food101N and Clothing1M. The code will be available at https://github.com/kaiwang960112/AFM.