论文标题
将深层神经网络抽象成概念级别的概念图,可解释性
Abstracting Deep Neural Networks into Concept Graphs for Concept Level Interpretability
论文作者
论文摘要
深度学习模型的黑盒本质使他们无法完全信任生物医学等领域。大多数解释性技术不会捕获人类所遵循的基于概念的推理。在这项工作中,我们试图通过构建对他们所学的概念的图形表示来了解训练有素的模型的行为,这些模型的行为是在医疗领域执行图像处理任务。在抽象上提取模型行为的图形表示,更高的概念水平将揭示这些模型的学习,并将帮助我们评估模型为预测所采取的步骤。我们显示了我们提出的实施在两个生物医学问题上的应用 - 脑肿瘤分割和眼底图像分类。我们通过制定上述概念级别图来提供模型的替代图形表示,这使得干预措施的问题更容易找到主动的推理跟踪。了解这些步道将提供对决策过程的层次结构的理解,然后是模型。 [以及模型的整体性质]。我们的框架可从https://github.com/koriavinash1/bioexp获得
The black-box nature of deep learning models prevents them from being completely trusted in domains like biomedicine. Most explainability techniques do not capture the concept-based reasoning that human beings follow. In this work, we attempt to understand the behavior of trained models that perform image processing tasks in the medical domain by building a graphical representation of the concepts they learn. Extracting such a graphical representation of the model's behavior on an abstract, higher conceptual level would unravel the learnings of these models and would help us to evaluate the steps taken by the model for predictions. We show the application of our proposed implementation on two biomedical problems - brain tumor segmentation and fundus image classification. We provide an alternative graphical representation of the model by formulating a concept level graph as discussed above, which makes the problem of intervention to find active inference trails more tractable. Understanding these trails would provide an understanding of the hierarchy of the decision-making process followed by the model. [As well as overall nature of model]. Our framework is available at https://github.com/koriavinash1/BioExp