论文标题

使用变质测试的深泡检测模型中的公平评估

Fairness Evaluation in Deepfake Detection Models using Metamorphic Testing

论文作者

Pu, Muxin, Kuan, Meng Yi, Lim, Nyee Thoang, Chong, Chun Yong, Lim, Mei Kuan

论文摘要

在存在异常情况下,深冰探测器的公平性没有得到很好的研究,尤其是如果这些异常在男性或女性受试者中更为突出。这项工作的主要动机是评估此类异常下的深泡检测模型的行为。但是,由于深度学习(DL)和人工智能(AI)系统的黑盒性质,当修改输入数据时,很难预测模型的性能。至关重要的是,如果无法正确解决此缺陷,它将对模型的公平性产生不利影响,并无意中歧视某些亚人口。因此,这项工作的目的是采用变质测试来检查选定的深击检测模型的可靠性,以及输入变化位置的转换如何影响输出。我们选择了MesoInception-4(一种最先进的深层检测模型)作为目标模型和化妆作为异常。化妆是通过利用DLIB库在填充RGB值之前获得68个面部标记的。变质关系是基于以下观点得出的,即输入图像的现实扰动,例如化妆,涉及眼线笔,眼影,腮红和唇膏(通常用于男性和女性图像的常见的化妆品外观),不应通过巨大的范围来改变模型的输出。此外,我们缩小了范围,以专注于揭示DL和AI系统中潜在的性别偏见。具体来说,我们有兴趣检查中含有4模型是否会产生不公平的决定,这应该是鲁棒性问题的结果。我们工作中的发现有可能为DL和AI系统中质量保证和公平性的新研究方向铺平道路。

Fairness of deepfake detectors in the presence of anomalies are not well investigated, especially if those anomalies are more prominent in either male or female subjects. The primary motivation for this work is to evaluate how deepfake detection model behaves under such anomalies. However, due to the black-box nature of deep learning (DL) and artificial intelligence (AI) systems, it is hard to predict the performance of a model when the input data is modified. Crucially, if this defect is not addressed properly, it will adversely affect the fairness of the model and result in discrimination of certain sub-population unintentionally. Therefore, the objective of this work is to adopt metamorphic testing to examine the reliability of the selected deepfake detection model, and how the transformation of input variation places influence on the output. We have chosen MesoInception-4, a state-of-the-art deepfake detection model, as the target model and makeup as the anomalies. Makeups are applied through utilizing the Dlib library to obtain the 68 facial landmarks prior to filling in the RGB values. Metamorphic relations are derived based on the notion that realistic perturbations of the input images, such as makeup, involving eyeliners, eyeshadows, blushes, and lipsticks (which are common cosmetic appearance) applied to male and female images, should not alter the output of the model by a huge margin. Furthermore, we narrow down the scope to focus on revealing potential gender biases in DL and AI systems. Specifically, we are interested to examine whether MesoInception-4 model produces unfair decisions, which should be considered as a consequence of robustness issues. The findings from our work have the potential to pave the way for new research directions in the quality assurance and fairness in DL and AI systems.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源