论文标题
公平感知的学习有偏见的自由代表
Fairness-Aware Learning with Prejudice Free Representations
论文作者
论文摘要
机器学习模型广泛用于做出对人类生活产生重大影响的决策。对这些模型进行了对历史数据的培训,这些数据可能包含有关敏感属性的信息,例如种族,性别,宗教等。这种敏感属性的存在可能会不公平地影响某些人群亚组。从数据中删除敏感特征是很简单的。但是,模型可能会从训练数据中可能存在的潜在敏感属性中获得偏见。这导致人们对所采用模型的公平性越来越担心。在本文中,我们提出了一种新型算法,该算法可以有效地识别和处理潜在的歧视特征。该方法是学习算法的不可知论者,可以很好地概括分类和回归任务。它也可以用作证明该模型在需要的情况下没有歧视法规依从性的关键帮助。该方法有助于收集无歧视功能,从而在确保模型的公平性的同时,可以改善模型性能。我们对公开可用现实世界数据集评估的实验结果表明,与其他方法相比,近乎理想的公平度量。
Machine learning models are extensively being used to make decisions that have a significant impact on human life. These models are trained over historical data that may contain information about sensitive attributes such as race, sex, religion, etc. The presence of such sensitive attributes can impact certain population subgroups unfairly. It is straightforward to remove sensitive features from the data; however, a model could pick up prejudice from latent sensitive attributes that may exist in the training data. This has led to the growing apprehension about the fairness of the employed models. In this paper, we propose a novel algorithm that can effectively identify and treat latent discriminating features. The approach is agnostic of the learning algorithm and generalizes well for classification as well as regression tasks. It can also be used as a key aid in proving that the model is free of discrimination towards regulatory compliance if the need arises. The approach helps to collect discrimination-free features that would improve the model performance while ensuring the fairness of the model. The experimental results from our evaluations on publicly available real-world datasets show a near-ideal fairness measurement in comparison to other methods.