论文标题
“如果没有发生,我为什么要改变决定?”:法官如何应对公共安全评估的反事实解释
"If it didn't happen, why would I change my decision?": How Judges Respond to Counterfactual Explanations for the Public Safety Assessment
论文作者
论文摘要
许多研究人员和政策制定者对算法解释感到兴奋,从而实现了更公平和负责任的决策。但是,最近的实验研究发现,解释并不总是改善人类对算法建议的使用。在这项研究中,我们阐明了人们如何解释和响应反事实解释(CFES) - 在审前风险评估工具(PRAIS)的背景下,模型的输出将如何随着其投入的边际变化而变化,这些解释表明了模型的产出将如何变化。我们与八名现任的美国州法院法官进行了思考审判,为他们提供了包括CFE在内的PRAI的建议。我们发现CFE并没有改变法官的决定。起初,法官误解了反事实是真实的 - 而不是假设的被告变化。一旦法官理解了反事实的含义,他们就会忽略他们,并说他们的作用只是为了对实际被告做出决定。法官还表达了一些理由,忽略或遵循没有CFE的Prai的建议。这些结果增加了文献,详细介绍了人们对算法和解释的反应的意外方式。他们还强调了与通过解释改善人类合作协作相关的新挑战。
Many researchers and policymakers have expressed excitement about algorithmic explanations enabling more fair and responsible decision-making. However, recent experimental studies have found that explanations do not always improve human use of algorithmic advice. In this study, we shed light on how people interpret and respond to counterfactual explanations (CFEs) -- explanations that show how a model's output would change with marginal changes to its input(s) -- in the context of pretrial risk assessment instruments (PRAIs). We ran think-aloud trials with eight sitting U.S. state court judges, providing them with recommendations from a PRAI that includes CFEs. We found that the CFEs did not alter the judges' decisions. At first, judges misinterpreted the counterfactuals as real -- rather than hypothetical -- changes to defendants. Once judges understood what the counterfactuals meant, they ignored them, stating their role is only to make decisions regarding the actual defendant in question. The judges also expressed a mix of reasons for ignoring or following the advice of the PRAI without CFEs. These results add to the literature detailing the unexpected ways in which people respond to algorithms and explanations. They also highlight new challenges associated with improving human-algorithm collaborations through explanations.