论文标题
头部和颈部肿瘤的自动分割:强大的变压器有多强?
Automatic Segmentation of Head and Neck Tumor: How Powerful Transformers Are?
论文作者
论文摘要
癌症是全球死亡的主要原因之一,头颈部(H&N)癌症是最普遍的类型之一。正电子发射断层扫描和计算机断层扫描用于检测,分段和量化肿瘤区域。在临床上,肿瘤分割非常耗时,容易出错。机器学习,尤其是深度学习可以帮助自动化此过程,从而产生与临床医生结果一样准确的结果。在本文中,我们研究了一种基于视觉变压器自动描述H&N肿瘤的方法,并将其结果与基于卷积神经网络(CNN)模型进行比较。我们使用来自CT和PET扫描的多模式数据来执行分割任务。我们表明,具有基于变压器的模型的解决方案具有与基于CNN的模型相当的结果。通过交叉验证,该模型的平均骰子相似性系数(DSC)为0.736,平均精度为0.766,平均召回率为0.766。就DSC分数而言,这仅比2020年竞赛获胜模型(在内部交叉验证)少0.021。在测试集中,该模型的性能类似,DSC为0.736,精度为0.773,召回0.760,在DSC中仅比2020年竞赛获胜模型低0.023。这项工作表明,通过基于变压器的模型进行癌症分割是一个有前途的研究领域,可以进一步探索。
Cancer is one of the leading causes of death worldwide, and head and neck (H&N) cancer is amongst the most prevalent types. Positron emission tomography and computed tomography are used to detect, segment and quantify the tumor region. Clinically, tumor segmentation is extensively time-consuming and prone to error. Machine learning, and deep learning in particular, can assist to automate this process, yielding results as accurate as the results of a clinician. In this paper, we investigate a vision transformer-based method to automatically delineate H&N tumor, and compare its results to leading convolutional neural network (CNN)-based models. We use multi-modal data from CT and PET scans to perform the segmentation task. We show that a solution with a transformer-based model has the potential to achieve comparable results to CNN-based ones. With cross validation, the model achieves a mean dice similarity coefficient (DSC) of 0.736, mean precision of 0.766 and mean recall of 0.766. This is only 0.021 less than the 2020 competition winning model (cross validated in-house) in terms of the DSC score. On the testing set, the model performs similarly, with DSC of 0.736, precision of 0.773, and recall of 0.760, which is only 0.023 lower in DSC than the 2020 competition winning model. This work shows that cancer segmentation via transformer-based models is a promising research area to further explore.