论文标题
揭示文本中的潜在偏见:方法和应用以同行评审
Uncovering Latent Biases in Text: Method and Application to Peer Review
论文作者
论文摘要
量化数值数量的系统差异,例如就业率和人口亚组之间的工资,为存在社会偏见的存在提供了令人信服的证据。但是,在为不同亚组成员(例如男性和非男性候选人的推荐信中)编写的文本中的偏见,尽管广泛报道了轶事,但仍具有挑战性。在这项工作中,我们介绍了一个新颖的框架,以量化由亚组成员指标的可见性引起的文本偏差。我们开发了非参数估计和推理程序来估计这一偏见。然后,我们正式制定了一种识别策略,以将估计的偏见与亚组成员资格指标的可见性联系起来,从而提供了从确定性隐藏策略更改之前和之后的时间段的观察结果。我们确定可以推断出“地面真相”偏见以评估我们的框架的应用,而不是依靠合成或辅助数据。具体来说,我们将框架应用于在会议采用双盲审查政策之前和之后,在著名的机器学习会议上的同行评审文本中量化偏见。我们在审查等级中显示了偏见的证据,这些评分是“地面真相”,并表明我们提出的框架准确地检测到了审查文本中的这些偏见,而无需访问审查评分。
Quantifying systematic disparities in numerical quantities such as employment rates and wages between population subgroups provides compelling evidence for the existence of societal biases. However, biases in the text written for members of different subgroups (such as in recommendation letters for male and non-male candidates), though widely reported anecdotally, remain challenging to quantify. In this work, we introduce a novel framework to quantify bias in text caused by the visibility of subgroup membership indicators. We develop a nonparametric estimation and inference procedure to estimate this bias. We then formalize an identification strategy to causally link the estimated bias to the visibility of subgroup membership indicators, provided observations from time periods both before and after an identity-hiding policy change. We identify an application wherein "ground truth" bias can be inferred to evaluate our framework, instead of relying on synthetic or secondary data. Specifically, we apply our framework to quantify biases in the text of peer reviews from a reputed machine learning conference before and after the conference adopted a double-blind reviewing policy. We show evidence of biases in the review ratings that serves as "ground truth", and show that our proposed framework accurately detects these biases from the review text without having access to the review ratings.