论文标题
副:机器学习模型的视觉反事实解释
ViCE: Visual Counterfactual Explanations for Machine Learning Models
论文作者
论文摘要
机器学习模型的预测准确性的持续提高允许其广泛的实际应用。但是,许多看似准确的模型做出的决定仍然需要域专家验证。此外,模型的最终用户还希望了解特定决策背后的原因。因此,对解释性的需求越来越重要。在本文中,我们提出了一种交互式视觉分析工具VICE,该工具生成了反事实解释,以将模型决策与情境化和评估。评估每个样本以确定翻转模型输出所需的最小变化集。这些解释旨在为最终用户提供个性化的可行见解,以了解或可能对自动化的决定进行竞争或改进。结果有效地显示在视觉界面中,其中突出显示了反事实说明,并为用户提供了交互式方法来探索数据和模型。该工具的功能通过其应用于家庭净值信用数据集的应用。
The continued improvements in the predictive accuracy of machine learning models have allowed for their widespread practical application. Yet, many decisions made with seemingly accurate models still require verification by domain experts. In addition, end-users of a model also want to understand the reasons behind specific decisions. Thus, the need for interpretability is increasingly paramount. In this paper we present an interactive visual analytics tool, ViCE, that generates counterfactual explanations to contextualize and evaluate model decisions. Each sample is assessed to identify the minimal set of changes needed to flip the model's output. These explanations aim to provide end-users with personalized actionable insights with which to understand, and possibly contest or improve, automated decisions. The results are effectively displayed in a visual interface where counterfactual explanations are highlighted and interactive methods are provided for users to explore the data and model. The functionality of the tool is demonstrated by its application to a home equity line of credit dataset.