论文标题
罗宾:一个可解释的精神分裂症诊断的可解释的深层网络
RobIn: A Robust Interpretable Deep Network for Schizophrenia Diagnosis
论文作者
论文摘要
精神分裂症是一种严重的心理健康状况,需要长期且复杂的诊断过程。但是,早期诊断对于控制症状至关重要。深度学习最近已成为分析和解释医疗数据的流行方式。过去将深度学习从大脑成像数据中诊断出来的尝试表现出了希望,但遭受了较大的训练申请差距 - 很难将实验室研究应用于现实世界。我们建议通过专注于易于访问的数据来减少这种训练应用差距。我们根据DSM-5标准收集患者的精神病观察数据集。由于使用DSM-5诊断精神分裂症的所有心理健康诊所已经记录了类似的数据,因此我们的方法很容易被整合到当前过程中,作为协助临床医生的工具,同时遵守正式的诊断标准。为了促进对系统的现实用法,我们表明它是可以解释和稳健的。了解机器学习工具如何进行诊断对于允许临床医生相信该诊断至关重要。为了解释框架,我们将两种互补的注意机制融合在一起,即“挤压和激发”和“自我注意力”,以分别确定全球属性的重要性和属性互动性。该模型使用这些重要性得分来做出决策。这使临床医生能够了解如何达到诊断,从而改善对模型的信任。由于机器学习模型通常很难从不同来源推广到数据,因此我们使用增强测试数据进行实验,以评估该模型对现实世界的适用性。我们发现我们的模型对扰动更为强大,因此应在临床环境中表现更好。通过10倍的交叉验证,它可以达到98%的精度。
Schizophrenia is a severe mental health condition that requires a long and complicated diagnostic process. However, early diagnosis is vital to control symptoms. Deep learning has recently become a popular way to analyse and interpret medical data. Past attempts to use deep learning for schizophrenia diagnosis from brain-imaging data have shown promise but suffer from a large training-application gap - it is difficult to apply lab research to the real world. We propose to reduce this training-application gap by focusing on readily accessible data. We collect a data set of psychiatric observations of patients based on DSM-5 criteria. Because similar data is already recorded in all mental health clinics that diagnose schizophrenia using DSM-5, our method could be easily integrated into current processes as a tool to assist clinicians, whilst abiding by formal diagnostic criteria. To facilitate real-world usage of our system, we show that it is interpretable and robust. Understanding how a machine learning tool reaches its diagnosis is essential to allow clinicians to trust that diagnosis. To interpret the framework, we fuse two complementary attention mechanisms, 'squeeze and excitation' and 'self-attention', to determine global attribute importance and attribute interactivity, respectively. The model uses these importance scores to make decisions. This allows clinicians to understand how a diagnosis was reached, improving trust in the model. Because machine learning models often struggle to generalise to data from different sources, we perform experiments with augmented test data to evaluate the model's applicability to the real world. We find that our model is more robust to perturbations, and should therefore perform better in a clinical setting. It achieves 98% accuracy with 10-fold cross-validation.