论文标题

费尔基特(Fairkit),费尔基特(Fairkit)在墙上,谁是最公平的?在培训公平模型中支持数据科学家

Fairkit, Fairkit, on the Wall, Who's the Fairest of Them All? Supporting Data Scientists in Training Fair Models

论文作者

Johnson, Brittany, Bartola, Jesse, Angell, Rico, Keith, Katherine, Witty, Sam, Giguere, Stephen J., Brun, Yuriy

论文摘要

现代软件在很大程度上依赖数据和机器学习,并影响塑造我们世界的决策。不幸的是,最近的研究表明,由于数据的偏见,软件系统经常将偏见注入他们的决策,从对男性声音的更好的封闭标题抄录而不是女性的声音,到为金融贷款的有色人种收费。为了解决机器学习的偏见,数据科学家需要工具,以帮助他们了解其特定数据域中模型质量和公平性之间的权衡。为此,我们提出了Fairkit-Learn,这是一种帮助数据科学家推理和理解公平性的工具包。 Fairkit-Learn与最先进的机器学习工具一起使用,并使用相同的接口来简化采用。它可以评估数千种通过多种机器学习算法,超参数和数据排列产生的模型,并计算和可视化帕累托最佳模型集,这些模型描述了公平性和质量之间最佳权衡。我们通过54名学生通过用户研究评估Fairkit-Learn,这表明使用Fairkit-Learn的学生与使用Scikit-Learn和IBM AI Fairness 360工具包的学生之间的模型相比,在公平和质量之间提供更好的平衡。借助Fairkit-Learn,用户可以选择比Scikit-Learn训练的型号高达67%,准确性高达67%。

Modern software relies heavily on data and machine learning, and affects decisions that shape our world. Unfortunately, recent studies have shown that because of biases in data, software systems frequently inject bias into their decisions, from producing better closed caption transcriptions of men's voices than of women's voices to overcharging people of color for financial loans. To address bias in machine learning, data scientists need tools that help them understand the trade-offs between model quality and fairness in their specific data domains. Toward that end, we present fairkit-learn, a toolkit for helping data scientists reason about and understand fairness. Fairkit-learn works with state-of-the-art machine learning tools and uses the same interfaces to ease adoption. It can evaluate thousands of models produced by multiple machine learning algorithms, hyperparameters, and data permutations, and compute and visualize a small Pareto-optimal set of models that describe the optimal trade-offs between fairness and quality. We evaluate fairkit-learn via a user study with 54 students, showing that students using fairkit-learn produce models that provide a better balance between fairness and quality than students using scikit-learn and IBM AI Fairness 360 toolkits. With fairkit-learn, users can select models that are up to 67% more fair and 10% more accurate than the models they are likely to train with scikit-learn.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源