论文标题
信息检索模型中的性别偏见
Debiasing Gender Bias in Information Retrieval Models
论文作者
论文摘要
文化,性别,种族等的偏见已经存在数十年,并影响了人类社会互动的许多领域。这些偏见已被证明会影响机器学习(ML)模型,并且对于自然语言处理(NLP),这可能会对下游任务产生严重的后果。减轻信息检索(IR)中的性别偏见对于避免传播刻板印象很重要。在这项工作中,我们采用了一个由两个组成部分组成的数据集:(1)文档与查询的相关性以及(2)文档的“性别”,其中代词被男性,女性和中性共轭取代。我们明确地表明,当对大型预训练的BERT编码器进行全面微调时,IR的预训练模型在零摄像检索任务中的性能不佳,并且使用适配器网络执行的轻量级微调可改善零拍摄的检索性能,几乎比基线高20%。我们还说明,预训练的模型具有性别偏见,导致检索到往往比女性更频繁的文章。我们通过引入一种偏见技术来克服这一点,该技术在模型更喜欢男性而不是女性时惩罚该模型,从而产生了一个有效的模型,该模型以平衡的方式检索文章。
Biases in culture, gender, ethnicity, etc. have existed for decades and have affected many areas of human social interaction. These biases have been shown to impact machine learning (ML) models, and for natural language processing (NLP), this can have severe consequences for downstream tasks. Mitigating gender bias in information retrieval (IR) is important to avoid propagating stereotypes. In this work, we employ a dataset consisting of two components: (1) relevance of a document to a query and (2) "gender" of a document, in which pronouns are replaced by male, female, and neutral conjugations. We definitively show that pre-trained models for IR do not perform well in zero-shot retrieval tasks when full fine-tuning of a large pre-trained BERT encoder is performed and that lightweight fine-tuning performed with adapter networks improves zero-shot retrieval performance almost by 20% over baseline. We also illustrate that pre-trained models have gender biases that result in retrieved articles tending to be more often male than female. We overcome this by introducing a debiasing technique that penalizes the model when it prefers males over females, resulting in an effective model that retrieves articles in a balanced fashion across genders.