论文标题
MISSU:通过自我依赖的Transunet进行3D医疗图像分割
MISSU: 3D Medical Image Segmentation via Self-distilling TransUNet
论文作者
论文摘要
U-NET在医学图像分割方面取得了巨大的成功。然而,它可能会遭受全球(远程)上下文相互作用和边缘维护的局限性。相比之下,变压器具有通过利用自我注意机制为编码器来捕获远程依赖性的出色能力。尽管变压器是为了建模对提取的特征图的远程依赖性而诞生的,但在处理高分辨率3D特征图时,它仍然具有极端的计算和空间复杂性。这促使我们设计有效的基于变压器的UNET模型,并研究基于变压器的网络体系结构对医疗图像分割任务的可行性。为此,我们建议为医学图像分割的基于变压器的UNET进行自我贡献,同时学习全局语义信息和本地空间详细信息。同时,首先提出了局部多尺度的融合块,以通过自我验证通过主CNN茎在编码器中的跳过连接中完善细节细节,仅在训练过程中计算出来,并在推断最小的开销时被删除。关于Brats 2019和混乱数据集的大量实验表明,我们的Missu在以前的最新方法中取得了最佳性能。代码和型号可在\ url {https://github.com/wangn123/missu.git}中获得
U-Nets have achieved tremendous success in medical image segmentation. Nevertheless, it may suffer limitations in global (long-range) contextual interactions and edge-detail preservation. In contrast, Transformer has an excellent ability to capture long-range dependencies by leveraging the self-attention mechanism into the encoder. Although Transformer was born to model the long-range dependency on the extracted feature maps, it still suffers from extreme computational and spatial complexities in processing high-resolution 3D feature maps. This motivates us to design the efficiently Transformer-based UNet model and study the feasibility of Transformer-based network architectures for medical image segmentation tasks. To this end, we propose to self-distill a Transformer-based UNet for medical image segmentation, which simultaneously learns global semantic information and local spatial-detailed features. Meanwhile, a local multi-scale fusion block is first proposed to refine fine-grained details from the skipped connections in the encoder by the main CNN stem through self-distillation, only computed during training and removed at inference with minimal overhead. Extensive experiments on BraTS 2019 and CHAOS datasets show that our MISSU achieves the best performance over previous state-of-the-art methods. Code and models are available at \url{https://github.com/wangn123/MISSU.git}