论文标题
一种模型是您所需要的:多任务学习启用同时组织学图像分割和分类
One Model is All You Need: Multi-Task Learning Enables Simultaneous Histology Image Segmentation and Classification
论文作者
论文摘要
数字化病理幻灯片的图像分析的最新性能激增很大程度上归因于深度学习的进步。深层模型可用于最初将各种结构定位在组织中,从而促进可解释的特征进行生物标志物发现。但是,这些模型通常是针对单个任务的训练,因此由于我们希望适应越来越多的不同任务的模型,因此扩展很差。同样,监督的深度学习模型非常饥饿,因此依靠大量的培训数据来表现良好。在本文中,我们提出了一种多任务学习方法,用于分割和分类核,腺体,Lumina和不同的组织区域,该方法利用了来自多个独立数据源的数据。在确保我们的任务与相同的组织类型和分辨率对齐的同时,我们可以通过一个网络同时进行有意义的同时预测。由于特征共享的结果,我们还表明,学习的表示形式可用于通过转移学习(包括核分类和标志环细胞检测)来提高其他任务的执行。作为这项工作的一部分,我们将开发的Cerberus模型培训大量数据,包括超过600K的分割对象和440K分类的补丁。我们使用我们的方法来处理来自TCGA的599个结直肠扫描图像,我们分别将3.77亿,900k和210万个核,腺体和Lumina进行了本地化,并使社区可用于下游分析。
The recent surge in performance for image analysis of digitised pathology slides can largely be attributed to the advances in deep learning. Deep models can be used to initially localise various structures in the tissue and hence facilitate the extraction of interpretable features for biomarker discovery. However, these models are typically trained for a single task and therefore scale poorly as we wish to adapt the model for an increasing number of different tasks. Also, supervised deep learning models are very data hungry and therefore rely on large amounts of training data to perform well. In this paper, we present a multi-task learning approach for segmentation and classification of nuclei, glands, lumina and different tissue regions that leverages data from multiple independent data sources. While ensuring that our tasks are aligned by the same tissue type and resolution, we enable meaningful simultaneous prediction with a single network. As a result of feature sharing, we also show that the learned representation can be used to improve the performance of additional tasks via transfer learning, including nuclear classification and signet ring cell detection. As part of this work, we train our developed Cerberus model on a huge amount of data, consisting of over 600K objects for segmentation and 440K patches for classification. We use our approach to process 599 colorectal whole-slide images from TCGA, where we localise 377 million, 900K and 2.1 million nuclei, glands and lumina, respectively and make the results available to the community for downstream analysis.