论文标题
PEG转移工作流识别挑战报告:多模式数据是否改善识别?
PEg TRAnsfer Workflow recognition challenge report: Does multi-modal data improve recognition?
论文作者
论文摘要
本文提出了“ PEG Transfort Workflow识别”(PETRAW)挑战的设计和结果,其目的是基于一种或几种模式在视频,运动学和细分数据中开发手术工作流识别方法,以研究其附加值。 PETRAW挑战提供了在虚拟模拟器上执行的150个PEG传输序列的数据集。该数据集由视频,运动学,语义分割和工作流注释组成,这些注释描述了三个不同粒度级别的序列:相位,步骤和活动。向参与者提出了五项任务:其中三个与对所有粒度的认可有关,其中一种可用的方式与所有粒度有关,而其他人则以多种方式解决了识别。使用平均应用依赖性平衡精度(AD准确性)用作评估度量标准,以考虑不平衡的类别,并且因为它在临床上比逐帧得分更相关。七个团队参加了至少一项任务,其中四个在所有任务中都参加了。通过使用视频和运动学数据,可以获得93%至90%的运动数据,从而获得了最佳结果。基于视频/运动学方法的方法与单模式的方法之间的改进对所有团队都显着。但是,必须考虑基于视频/运动学和基于运动学的方法之间测试执行时间的差异。花费不到3%的改进的计算时间增加20到200倍是否相关? PETRAW数据集可在www.synapse.org/petraw上公开获得,以鼓励进一步研究手术工作流识别。
This paper presents the design and results of the "PEg TRAnsfert Workflow recognition" (PETRAW) challenge whose objective was to develop surgical workflow recognition methods based on one or several modalities, among video, kinematic, and segmentation data, in order to study their added value. The PETRAW challenge provided a data set of 150 peg transfer sequences performed on a virtual simulator. This data set was composed of videos, kinematics, semantic segmentation, and workflow annotations which described the sequences at three different granularity levels: phase, step, and activity. Five tasks were proposed to the participants: three of them were related to the recognition of all granularities with one of the available modalities, while the others addressed the recognition with a combination of modalities. Average application-dependent balanced accuracy (AD-Accuracy) was used as evaluation metric to take unbalanced classes into account and because it is more clinically relevant than a frame-by-frame score. Seven teams participated in at least one task and four of them in all tasks. Best results are obtained with the use of the video and the kinematics data with an AD-Accuracy between 93% and 90% for the four teams who participated in all tasks. The improvement between video/kinematic-based methods and the uni-modality ones was significant for all of the teams. However, the difference in testing execution time between the video/kinematic-based and the kinematic-based methods has to be taken into consideration. Is it relevant to spend 20 to 200 times more computing time for less than 3% of improvement? The PETRAW data set is publicly available at www.synapse.org/PETRAW to encourage further research in surgical workflow recognition.