论文标题
基于Swin-Unet和多模式图像的腮腺MRI分割
Parotid Gland MRI Segmentation Based on Swin-Unet and Multimodal Images
论文作者
论文摘要
背景和客观:腮腺肿瘤约占头颈肿瘤的2%至10%。术前肿瘤定位,鉴别诊断以及随后选择适当的腮腺肿瘤治疗方法。然而,这些肿瘤的相对稀有性和高度分散的组织类型使基于术前放射线学对这种肿瘤病变的微妙鉴别诊断造成了未满足的需求。最近,深度学习方法发展迅速,尤其是变形金刚在计算机视觉中击败了传统的卷积神经网络。为计算机视觉任务提出了许多新的基于变压器的网络。方法:在这项研究中,收集了多中心多模束MR图像。使用了基于变压器的SWIN-UNET。短时间反转恢复,T1加权和T2加权方式的MR图像合并为三通道数据,以训练网络。我们实现了对腮腺和肿瘤感兴趣区域的分割。结果:模型在测试集上的骰子相似系数为88.63%,平均像素精度为99.31%,平均交叉点为83.99%,而Hausdorff的距离为3.04。然后在本文中设计了一系列比较实验,以进一步验证算法的分割性能。结论:实验结果表明,我们的方法对腮腺和肿瘤分割具有良好的结果。基于变压器的网络在医学图像领域优于传统的卷积神经网络。
Background and objective: Parotid gland tumors account for approximately 2% to 10% of head and neck tumors. Preoperative tumor localization, differential diagnosis, and subsequent selection of appropriate treatment for parotid gland tumors are critical. However, the relative rarity of these tumors and the highly dispersed tissue types have left an unmet need for a subtle differential diagnosis of such neoplastic lesions based on preoperative radiomics. Recently, deep learning methods have developed rapidly, especially Transformer beats the traditional convolutional neural network in computer vision. Many new Transformer-based networks have been proposed for computer vision tasks. Methods: In this study, multicenter multimodal parotid gland MR images were collected. The Swin-Unet which was based on Transformer was used. MR images of short time inversion recovery, T1-weighted and T2-weighted modalities were combined into three-channel data to train the network. We achieved segmentation of the region of interest for parotid gland and tumor. Results: The Dice-Similarity Coefficient of the model on the test set was 88.63%, Mean Pixel Accuracy was 99.31%, Mean Intersection over Union was 83.99%, and Hausdorff Distance was 3.04. Then a series of comparison experiments were designed in this paper to further validate the segmentation performance of the algorithm. Conclusions: Experimental results showed that our method has good results for parotid gland and tumor segmentation. The Transformer-based network outperforms the traditional convolutional neural network in the field of medical images.