论文标题
公平的对手网络
Fair Adversarial Networks
论文作者
论文摘要
人类判断的影响在整个分析行业的数据集中无处不在,但是人们却被众所周知,这是容易发生各种偏见的亚最佳决策者。然后分析有偏见的数据集会导致分析的偏差。受保护特征(例如种族)的偏见特别令人感兴趣,因为它不仅可能使分析过程的输出次优最佳,而且是非法的。通过将分析结果限制为公平,反对偏见是有问题的,因为a)公平性缺乏普遍接受的定义,而同时有些定义是相互排斥的,b)使用优化约束来确保公平性与大多数分析管道不符。这两个问题都是通过从数据中消除偏差并返回更改数据集的方法来解决的。这种方法的目的不仅旨在删除实际的偏差变量(例如竞赛),还旨在更改所有代理变量(例如邮政编码),因此无法从其他数据中检测到偏差变量。使用这种方法的优点是,公平性的定义是数据中缺乏可检测到的偏差(而不是分析的输出)是普遍的,因此解决了问题(a)。此外,随着数据的更改以消除偏见,问题(b)消失了,因为分析管道可能保持不变。这种方法已被多种技术解决方案采用。但是,在消除多元,非线性和非二进制偏见的能力方面,它们似乎都不令人满意。因此,在本文中,我提出了公平对抗网络的概念,作为一种易于实现的通用方法,可从数据中消除偏见。本文表明,公平的对抗网络实现了这一目标。
The influence of human judgement is ubiquitous in datasets used across the analytics industry, yet humans are known to be sub-optimal decision makers prone to various biases. Analysing biased datasets then leads to biased outcomes of the analysis. Bias by protected characteristics (e.g. race) is of particular interest as it may not only make the output of analytical process sub-optimal, but also illegal. Countering the bias by constraining the analytical outcomes to be fair is problematic because A) fairness lacks a universally accepted definition, while at the same time some definitions are mutually exclusive, and B) the use of optimisation constraints ensuring fairness is incompatible with most analytical pipelines. Both problems are solved by methods which remove bias from the data and returning an altered dataset. This approach aims to not only remove the actual bias variable (e.g. race), but also alter all proxy variables (e.g. postcode) so the bias variable is not detectable from the rest of the data. The advantage of using this approach is that the definition of fairness as a lack of detectable bias in the data (as opposed to the output of analysis) is universal and therefore solves problem (A). Furthermore, as the data is altered to remove bias the problem (B) disappears because the analytical pipelines can remain unchanged. This approach has been adopted by several technical solutions. None of them, however, seems to be satisfactory in terms of ability to remove multivariate, non-linear and non-binary biases. Therefore, in this paper I propose the concept of Fair Adversarial Networks as an easy-to-implement general method for removing bias from data. This paper demonstrates that Fair Adversarial Networks achieve this aim.