论文标题

一票否决:低光青光眼诊断的半监督学习

One-Vote Veto: Semi-Supervised Learning for Low-Shot Glaucoma Diagnosis

论文作者

Fan, Rui, Bowd, Christopher, Brye, Nicole, Christopher, Mark, Weinreb, Robert N., Kriegman, David, Zangwill, Linda M.

论文摘要

卷积神经网络(CNN)是一种从眼睛图像中自动化的青光眼诊断的有前途的技术,并且这些图像通常作为眼科检查的一部分获取。然而,CNN通常需要大量标记的数据进行培训,这在许多生物医学图像分类应用中可能无法使用,尤其是在疾病很少见,专家标记的情况下是昂贵的。本文为解决这个问题做出了两种贡献:(1)它扩展了传统的暹罗网络,并在标记的数据受到限制和不平衡时介绍了一种用于低速学习的培训方法,并且(2)它介绍了一种新型的半监督学习策略,该策略使用其他未标记的培训数据来实现更准确的准确性。我们提出的多任务暹罗网络(MTSN)可以采用任何骨干CNN,我们用四个骨干CNN证明其准确性在有限的培训数据中可以使用培训的骨干CNN的准确性,该骨干CNN经过培训的数据集,该数据集大于50倍。我们还介绍了一个专门为MTSN设计的半监督学习策略(一种半监督的学习策略)。通过考虑对未标记的培训数据的自我预测和对比预测,OVV自我培训提供了额外的伪标签,以微调预训练的MTSN。通过使用15年来获得66,715个眼底照片的大型(不平衡)数据集,广泛的实验结果证明了使用MTSN和半手不见的学习和半监督学习的有效性和OVV自我训练。在不同条件下(相机,工具,位置,种群)中获得的三个额外的较小的临床数据集来证明所提出的方法的普遍性。

Convolutional neural networks (CNNs) are a promising technique for automated glaucoma diagnosis from images of the fundus, and these images are routinely acquired as part of an ophthalmic exam. Nevertheless, CNNs typically require a large amount of well-labeled data for training, which may not be available in many biomedical image classification applications, especially when diseases are rare and where labeling by experts is costly. This article makes two contributions to address this issue: (1) It extends the conventional Siamese network and introduces a training method for low-shot learning when labeled data are limited and imbalanced, and (2) it introduces a novel semi-supervised learning strategy that uses additional unlabeled training data to achieve greater accuracy. Our proposed multi-task Siamese network (MTSN) can employ any backbone CNN, and we demonstrate with four backbone CNNs that its accuracy with limited training data approaches the accuracy of backbone CNNs trained with a dataset that is 50 times larger. We also introduce One-Vote Veto (OVV) self-training, a semi-supervised learning strategy that is designed specifically for MTSNs. By taking both self-predictions and contrastive predictions of the unlabeled training data into account, OVV self-training provides additional pseudo labels for fine-tuning a pre-trained MTSN. Using a large (imbalanced) dataset with 66,715 fundus photographs acquired over 15 years, extensive experimental results demonstrate the effectiveness of low-shot learning with MTSN and semi-supervised learning with OVV self-training. Three additional, smaller clinical datasets of fundus images acquired under different conditions (cameras, instruments, locations, populations) are used to demonstrate the generalizability of the proposed methods.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源