论文标题
在人类决策中的解释,公平和适当的依赖
Explanations, Fairness, and Appropriate Reliance in Human-AI Decision-Making
论文作者
论文摘要
在这项工作中,我们研究了基于特征的解释对AI辅助决策的分配公平性的影响,特别是专注于预测短文本BIOS职业的任务。我们还研究了任何效果是如何由人类的公平感知及其对AI建议的依赖。我们的发现表明,解释会影响公平感,这反过来又与人类遵守AI建议的倾向有关。但是,我们看到这种解释并不能使人类辨别正确和错误的AI建议。相反,我们表明它们可能会影响依赖性,而不论AI建议的正确性。根据哪些具有解释的重点,这可以促进或阻碍分布式公平:当解释突出显示任务 - irrelevant并显然与敏感属性相关的功能时,这会促使其反对与性别刻板印象相符的AI建议。同时,如果解释出现与任务相关,则会引起依赖刻板印象对准错误的依赖行为。这些结果表明,基于特征的解释不是提高分配公平性的可靠机制。
In this work, we study the effects of feature-based explanations on distributive fairness of AI-assisted decisions, specifically focusing on the task of predicting occupations from short textual bios. We also investigate how any effects are mediated by humans' fairness perceptions and their reliance on AI recommendations. Our findings show that explanations influence fairness perceptions, which, in turn, relate to humans' tendency to adhere to AI recommendations. However, we see that such explanations do not enable humans to discern correct and incorrect AI recommendations. Instead, we show that they may affect reliance irrespective of the correctness of AI recommendations. Depending on which features an explanation highlights, this can foster or hinder distributive fairness: when explanations highlight features that are task-irrelevant and evidently associated with the sensitive attribute, this prompts overrides that counter AI recommendations that align with gender stereotypes. Meanwhile, if explanations appear task-relevant, this induces reliance behavior that reinforces stereotype-aligned errors. These results imply that feature-based explanations are not a reliable mechanism to improve distributive fairness.