论文标题
对可解释特征的需求:动机和分类学
The Need for Interpretable Features: Motivation and Taxonomy
论文作者
论文摘要
通过为现实世界域开发和解释机器学习(ML)应用的丰富经验,我们了解到ML模型仅与它们的功能一样可解释。即使是简单,高度可解释的模型类型,例如回归模型,如果使用不可解释的功能,也很难理解。不同的用户,尤其是那些使用ML模型在其域中决策的用户可能需要不同级别和特征的解释性。此外,根据我们的经验,我们声称“可解释的功能”一词不是具体的,也不足够详细,无法捕获影响ML解释的有用性的全部程度。在本文中,我们激励并讨论了三个关键的课程:1)应该更多地关注我们所谓的可解释的特征空间或对采取现实世界行动的领域专家有用的特征状态,2)需要对这些领域可能需要的特征属性进行形式的分类特性,这些特征是这些领域所要求的,这些域名是在本文中既可以享用的,又是在此论文中概述的模型,以及3)的模型,以及3) - 与3)相同的模型 - 与该领域相同的模型,请参阅该领域的模型,请参阅3),这是一个模型,请参与其中的模型,请参与其中的模型,请参与其中的模型,这是一定的。 ML转换为模型准备功能。
Through extensive experience developing and explaining machine learning (ML) applications for real-world domains, we have learned that ML models are only as interpretable as their features. Even simple, highly interpretable model types such as regression models can be difficult or impossible to understand if they use uninterpretable features. Different users, especially those using ML models for decision-making in their domains, may require different levels and types of feature interpretability. Furthermore, based on our experiences, we claim that the term "interpretable feature" is not specific nor detailed enough to capture the full extent to which features impact the usefulness of ML explanations. In this paper, we motivate and discuss three key lessons: 1) more attention should be given to what we refer to as the interpretable feature space, or the state of features that are useful to domain experts taking real-world actions, 2) a formal taxonomy is needed of the feature properties that may be required by these domain experts (we propose a partial taxonomy in this paper), and 3) transforms that take data from the model-ready state to an interpretable form are just as essential as traditional ML transforms that prepare features for the model.