论文标题

对文本复杂推理的负责和可解释的方法

Accountable and Explainable Methods for Complex Reasoning over Text

论文作者

Atanasova, Pepa

论文摘要

机器学习(ML)模型的主要关注点是它们的不透明度。它们被部署在越来越多的应用程序中,在这些应用程序中,它们通常作为黑匣子运行,而这些应用程序无法提供预测的解释。除其他外,与对模型原理缺乏理解有关的潜在危害包括侵犯隐私,对抗性操纵和不公平的歧视。结果,政策和法律,哲学和计算机科学的作品中,ML模型的问责制和透明度已被认为是关键的逃亡者。 在计算机科学中,通过开发问责制和透明度方法研究了ML模型的决策过程。问责方法,例如对抗性攻击和诊断数据集,暴露了ML模型的漏洞,这些漏洞可能导致其预测中导致恶意操纵或系统性故障。透明度方法解释了模型预测背后的理由,从而获得了相关利益相关者的信任,并可能揭示了模型决策中的错误和不公平性。为此,透明度方法也必须满足问责制要求,例如,对模型的基本原理非常强大和忠实。 本论文提出了我的研究,该研究扩大了我们在为文本上复杂的推理任务开发的ML模型领域和透明度领域的集体知识。

A major concern of Machine Learning (ML) models is their opacity. They are deployed in an increasing number of applications where they often operate as black boxes that do not provide explanations for their predictions. Among others, the potential harms associated with the lack of understanding of the models' rationales include privacy violations, adversarial manipulations, and unfair discrimination. As a result, the accountability and transparency of ML models have been posed as critical desiderata by works in policy and law, philosophy, and computer science. In computer science, the decision-making process of ML models has been studied by developing accountability and transparency methods. Accountability methods, such as adversarial attacks and diagnostic datasets, expose vulnerabilities of ML models that could lead to malicious manipulations or systematic faults in their predictions. Transparency methods explain the rationales behind models' predictions gaining the trust of relevant stakeholders and potentially uncovering mistakes and unfairness in models' decisions. To this end, transparency methods have to meet accountability requirements as well, e.g., being robust and faithful to the underlying rationales of a model. This thesis presents my research that expands our collective knowledge in the areas of accountability and transparency of ML models developed for complex reasoning tasks over text.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源