论文标题
及时调整参数有效的医学图像分段
Prompt Tuning for Parameter-efficient Medical Image Segmentation
论文作者
论文摘要
在自学计划中预先培训的神经网络在具有稀缺注释的数据丰富环境中运行时已成为标准。因此,以参数效率但有效的方式将模型微调到下游任务,例如对于在语义细分的情况下,对于新的类别来说,重要性越来越重要。在这项工作中,我们提出并研究了几项贡献,以实现两个医学成像数据集上的语义分割的参数效率但有效的适应。依靠最近普及的及时调整方法,我们提供了一种及时的UNET(PUNET)体系结构,该体系结构是在预训练后冷冻的,但可以通过依赖类的可学习提示令牌在整个网络中适应。我们基于对在线生成的原型(对比的原型分配,CPA)的分配,以专门的密集自我统计方案进行培训。我们证明,所得的神经网络模型能够减轻CT成像数据集上完全微调和有效调整模型之间的差距。因此,TCIA/BTCV数据集的全面微调和及时调整变体之间的差异仅为3.83 pp,在平均骰子相似度系数(DSC)中,CT-ORG数据集的差异仅为2.67 pp,仅在%中,而仅提示与预先验证的后台6.85的参数相对应的是0.85%,均为0.85%。这项工作的代码可在https://github.com/marcdcfischer/punet上获得。
Neural networks pre-trained on a self-supervision scheme have become the standard when operating in data rich environments with scarce annotations. As such, fine-tuning a model to a downstream task in a parameter-efficient but effective way, e.g. for a new set of classes in the case of semantic segmentation, is of increasing importance. In this work, we propose and investigate several contributions to achieve a parameter-efficient but effective adaptation for semantic segmentation on two medical imaging datasets. Relying on the recently popularized prompt tuning approach, we provide a prompt-able UNet (PUNet) architecture, that is frozen after pre-training, but adaptable throughout the network by class-dependent learnable prompt tokens. We pre-train this architecture with a dedicated dense self-supervision scheme based on assignments to online generated prototypes (contrastive prototype assignment, CPA) of a student teacher combination alongside a concurrent segmentation loss on a subset of classes. We demonstrate that the resulting neural network model is able to attenuate the gap between fully fine-tuned and parameter-efficiently adapted models on CT imaging datasets. As such, the difference between fully fine-tuned and prompt-tuned variants amounts to only 3.83 pp for the TCIA/BTCV dataset and 2.67 pp for the CT-ORG dataset in the mean Dice Similarity Coefficient (DSC, in %) while only prompt tokens, corresponding to 0.85% of the pre-trained backbone model with 6.8M frozen parameters, are adjusted. The code for this work is available on https://github.com/marcdcfischer/PUNet .