论文标题
使用参数积分概率度量的学习公平表示
Learning fair representation with a parametric integral probability metric
论文作者
论文摘要
由于它们对社会决策具有重要影响,因此AI算法不仅应该是准确的,而且应该是公平的。在公平性AI的各种算法中,学习公平代表(LFR)的目标是在诸如性别和种族等敏感变量方面找到公平的代表,并受到了很多关注。对于LFR,对抗训练方案通常像生成对抗网络类型算法一样使用。但是,歧视者的选择是在没有理由的情况下进行的。在本文中,我们为LFR提出了一种新的对抗性训练计划,其中使用具有特定参数歧视剂家族的积分概率度量(IPM)。提出的LFR算法的最显着结果是其关于最终预测模型公平性的理论保证,尚未考虑。也就是说,我们在表示的公平性与构建在表示顶部的预测模型的公平性之间(即将表示形式用作输入)之间得出了理论关系。此外,通过数值实验,我们表明我们提出的LFR算法在计算上更轻且更稳定,并且最终的预测模型具有竞争性或优于其他LFR算法,使用更复杂的歧视器。
As they have a vital effect on social decision-making, AI algorithms should be not only accurate but also fair. Among various algorithms for fairness AI, learning fair representation (LFR), whose goal is to find a fair representation with respect to sensitive variables such as gender and race, has received much attention. For LFR, the adversarial training scheme is popularly employed as is done in the generative adversarial network type algorithms. The choice of a discriminator, however, is done heuristically without justification. In this paper, we propose a new adversarial training scheme for LFR, where the integral probability metric (IPM) with a specific parametric family of discriminators is used. The most notable result of the proposed LFR algorithm is its theoretical guarantee about the fairness of the final prediction model, which has not been considered yet. That is, we derive theoretical relations between the fairness of representation and the fairness of the prediction model built on the top of the representation (i.e., using the representation as the input). Moreover, by numerical experiments, we show that our proposed LFR algorithm is computationally lighter and more stable, and the final prediction model is competitive or superior to other LFR algorithms using more complex discriminators.