论文标题
有条件的监督对比度学习,用于公平文本分类
Conditional Supervised Contrastive Learning for Fair Text Classification
论文作者
论文摘要
对比表示学习因其在图像和顺序数据中的学习表示方面的出色表现而引起了很多关注。但是,学习的表示可能会导致下游任务中的性能差异,例如增加代表性不足的毒性评论分类的沉默。鉴于这项挑战,在这项工作中,我们研究了学习公平的表现,这些表现能够满足通过对比度学习的公平概念,称为文本分类的均等赔率。具体而言,我们首先通过公平限制和有条件监督的对比目标分析学习表示之间的联系,然后建议使用条件监督的对比目标来学习文本分类的公平表示。我们在两个文本数据集上进行实验,以证明我们的方法在平衡任务绩效和偏置缓解基线之间的权衡方面的有效性,以进行文本分类。此外,我们还表明所提出的方法在不同的高参数设置中是稳定的。
Contrastive representation learning has gained much attention due to its superior performance in learning representations from both image and sequential data. However, the learned representations could potentially lead to performance disparities in downstream tasks, such as increased silencing of underrepresented groups in toxicity comment classification. In light of this challenge, in this work, we study learning fair representations that satisfy a notion of fairness known as equalized odds for text classification via contrastive learning. Specifically, we first theoretically analyze the connections between learning representations with a fairness constraint and conditional supervised contrastive objectives, and then propose to use conditional supervised contrastive objectives to learn fair representations for text classification. We conduct experiments on two text datasets to demonstrate the effectiveness of our approaches in balancing the trade-offs between task performance and bias mitigation among existing baselines for text classification. Furthermore, we also show that the proposed methods are stable in different hyperparameter settings.