论文标题
演示样式对算法偏见的人体检测的影响
The Impact of Presentation Style on Human-In-The-Loop Detection of Algorithmic Bias
论文作者
论文摘要
尽管决策者已经开始采用机器学习,但机器学习模型可能会对某些人口统计组进行偏见。半自动化的偏置检测工具通常会使用建议列表或视觉提示呈现自动检测偏差的报告。但是,缺乏有关在哪种情况下使用哪种演示样式的指导。我们对16名参与者进行了一项小型实验室研究,以研究演示方式如何影响用户审查偏见报告的行为。参与者既使用具有推荐列表的原型,也使用了带有视觉提示的原型。我们发现,参与者通常想研究未自动检测到偏见的绩效指标。但是,当将原型与推荐列表一起使用时,它们往往会更少考虑此类措施。基于调查结果,我们提出了信息负载和综合性,作为表征偏见检测任务的两个轴,并说明了如何采用两个轴来推理何时使用建议列表或视觉提示。
While decision makers have begun to employ machine learning, machine learning models may make predictions that bias against certain demographic groups. Semi-automated bias detection tools often present reports of automatically-detected biases using a recommendation list or visual cues. However, there is a lack of guidance concerning which presentation style to use in what scenarios. We conducted a small lab study with 16 participants to investigate how presentation style might affect user behaviors in reviewing bias reports. Participants used both a prototype with a recommendation list and a prototype with visual cues for bias detection. We found that participants often wanted to investigate the performance measures that were not automatically detected as biases. Yet, when using the prototype with a recommendation list, they tended to give less consideration to such measures. Grounded in the findings, we propose information load and comprehensiveness as two axes for characterizing bias detection tasks and illustrate how the two axes could be adopted to reason about when to use a recommendation list or visual cues.