论文标题

公平意识的在线个性化

Fairness-Aware Online Personalization

论文作者

Lal, G Roshan, Geyik, Sahin Cem, Kenthapadi, Krishnaram

论文摘要

诸如贷款,招聘和大学录取等关键应用中的决策已经见证了算法模型和技术的使用越来越多,这是由于多种因素汇合而导致的,例如普遍存在的连通性,收集,聚集和处理大量的使用云计算的精细数据,以及访问精致的机器学习模型。这些应用程序通常是由搜索和推荐系统提供动力的,而搜索和推荐系统又使用个性化排名算法。同时,人们对使用此类数据驱动系统带来的道德和法律挑战的认识越来越高。来自不同学科的研究人员和从业人员最近强调了此类系统对某些人群群体进行区分的潜力,这是由于用于学习其基本建议模型的数据集中的偏见。我们介绍了涉及个人排名的在线个性化设置中的公平性研究。从一个公平的启动机器学习模型开始,我们首先证明在线个性化可能会导致该模型学会以不公平的方式行事,如果用户对他/她的回答有偏见。为此,我们构建了一个风格化的模型,用于生成具有潜在偏见的特征以及潜在偏见的标签的训练数据,并量化模型在用户以偏见方式响应的偏见程度,例如在许多现实世界情景中。然后,我们在公平限制下提出学习个性化模型的问题,并提出一种基于正则化的方法来减轻机器学习中的偏见。我们通过具有不同参数设置的广泛模拟来证明方法的功效。代码:https://github.com/groshanlal/fairness-aware-online-personalization

Decision making in crucial applications such as lending, hiring, and college admissions has witnessed increasing use of algorithmic models and techniques as a result of a confluence of factors such as ubiquitous connectivity, ability to collect, aggregate, and process large amounts of fine-grained data using cloud computing, and ease of access to applying sophisticated machine learning models. Quite often, such applications are powered by search and recommendation systems, which in turn make use of personalized ranking algorithms. At the same time, there is increasing awareness about the ethical and legal challenges posed by the use of such data-driven systems. Researchers and practitioners from different disciplines have recently highlighted the potential for such systems to discriminate against certain population groups, due to biases in the datasets utilized for learning their underlying recommendation models. We present a study of fairness in online personalization settings involving the ranking of individuals. Starting from a fair warm-start machine-learned model, we first demonstrate that online personalization can cause the model to learn to act in an unfair manner if the user is biased in his/her responses. For this purpose, we construct a stylized model for generating training data with potentially biased features as well as potentially biased labels and quantify the extent of bias that is learned by the model when the user responds in a biased manner as in many real-world scenarios. We then formulate the problem of learning personalized models under fairness constraints and present a regularization based approach for mitigating biases in machine learning. We demonstrate the efficacy of our approach through extensive simulations with different parameter settings. Code: https://github.com/groshanlal/Fairness-Aware-Online-Personalization

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源