论文标题
通过相似性的统计量度学习解开文本表示
Learning Disentangled Textual Representations via Statistical Measures of Similarity
论文作者
论文摘要
在使用文本数据时,自然应用分开表示是公平的分类,其目标是在数据中可能存在的敏感属性(例如,年龄,性别或种族)进行预测而不会受到偏见(或影响)。从文本表示中解开敏感属性的主要方法依赖于同时学习的惩罚术语,涉及对抗性损失(例如,歧视者)或信息度量(例如,相互信息)。但是,这些方法需要培训深层神经网络,并为表示模型的每个更新进行多个参数更新。实际上,所得的嵌套优化循环既耗时又耗时,从而增加了优化动态的复杂性,并且需要精细的超参数选择(例如,学习率,体系结构)。在这项工作中,我们介绍了一个正规机构的家庭,以学习不需要培训的删除表示的表示。这些正规化器基于条件概率分布相对于敏感属性之间相似性的统计量度。我们的新型正规化器不需要额外的培训,更快,并且不涉及额外的调整,同时与预审预认证和随机初始化的文本编码器相结合时,取得了更好的结果。
When working with textual data, a natural application of disentangled representations is fair classification where the goal is to make predictions without being biased (or influenced) by sensitive attributes that may be present in the data (e.g., age, gender or race). Dominant approaches to disentangle a sensitive attribute from textual representations rely on learning simultaneously a penalization term that involves either an adversarial loss (e.g., a discriminator) or an information measure (e.g., mutual information). However, these methods require the training of a deep neural network with several parameter updates for each update of the representation model. As a matter of fact, the resulting nested optimization loop is both time consuming, adding complexity to the optimization dynamic, and requires a fine hyperparameter selection (e.g., learning rates, architecture). In this work, we introduce a family of regularizers for learning disentangled representations that do not require training. These regularizers are based on statistical measures of similarity between the conditional probability distributions with respect to the sensitive attributes. Our novel regularizers do not require additional training, are faster and do not involve additional tuning while achieving better results both when combined with pretrained and randomly initialized text encoders.