论文标题
使用变压器转换医学成像?对关键属性,当前进展和未来观点的比较综述
Transforming medical imaging with Transformers? A comparative review of key properties, current progresses, and future perspectives
论文作者
论文摘要
Transformer是深度学习的最新技术进步,它在自然语言处理或计算机视觉方面已经普遍存在。由于医学成像与计算机视觉有一定的相似之处,因此很自然地询问医疗成像中变压器的现状并提出问题:变压器模型可以转换医学成像吗?在本文中,我们试图对询问做出回应。 After a brief introduction of the fundamentals of Transformers, especially in comparison with convolutional neural networks (CNNs), and highlighting key defining properties that characterize the Transformers, we offer a comprehensive review of the state-of-the-art Transformer-based approaches for medical imaging and exhibit current research progresses made in the areas of medical image segmentation, recognition, detection, registration, reconstruction, enhancement, etc. In particular, what distinguishes our review在其组织中,基于变压器的密钥定义属性,这些属性主要来自比较变压器和CNN及其类型的体系结构的类型,这些构建结构指定了变压器和CNN合并的方式,所有这些方式都可以帮助读者最好地了解审查方法背后的理由。我们以讨论未来观点的讨论结束。
Transformer, the latest technological advance of deep learning, has gained prevalence in natural language processing or computer vision. Since medical imaging bear some resemblance to computer vision, it is natural to inquire about the status quo of Transformers in medical imaging and ask the question: can the Transformer models transform medical imaging? In this paper, we attempt to make a response to the inquiry. After a brief introduction of the fundamentals of Transformers, especially in comparison with convolutional neural networks (CNNs), and highlighting key defining properties that characterize the Transformers, we offer a comprehensive review of the state-of-the-art Transformer-based approaches for medical imaging and exhibit current research progresses made in the areas of medical image segmentation, recognition, detection, registration, reconstruction, enhancement, etc. In particular, what distinguishes our review lies in its organization based on the Transformer's key defining properties, which are mostly derived from comparing the Transformer and CNN, and its type of architecture, which specifies the manner in which the Transformer and CNN are combined, all helping the readers to best understand the rationale behind the reviewed approaches. We conclude with discussions of future perspectives.