论文标题
卷积神经网络模型的功能可视化,该模型接受了神经影像学数据的培训
Feature visualization for convolutional neural network models trained on neuroimaging data
论文作者
论文摘要
在临床决策中应用机器学习模型的主要先决条件是信任和解释性。神经影像学界的当前解释性研究主要集中在解释训练有素的模型的个人决策,例如通过卷积神经网络(CNN)获得。可以创建使用归因方法,例如层次相关性传播或Shap Heatmap,以突出显示哪些输入区域与决策更相关。尽管这允许检测潜在的数据集偏见,并且可以用作人类专家的指南,但它不允许了解该模型所学的基本原理。在这项研究中,我们取而代之的是,据我们所知,首次使用神经成像CNN的特征可视化结果。特别是,我们已经培训了CNN,用于基于结构磁共振成像(MRI)数据的不同任务,包括性别分类和人工病变分类。然后,我们已经迭代生成的图像可以最大程度地激活特定的神经元,以可视化它们响应的模式。为了改善可视化,我们比较了几种正则化策略。由此产生的图像揭示了人造病变的学习概念,包括其形状,但对于性别分类任务中的抽象特征仍然很难解释。
A major prerequisite for the application of machine learning models in clinical decision making is trust and interpretability. Current explainability studies in the neuroimaging community have mostly focused on explaining individual decisions of trained models, e.g. obtained by a convolutional neural network (CNN). Using attribution methods such as layer-wise relevance propagation or SHAP heatmaps can be created that highlight which regions of an input are more relevant for the decision than others. While this allows the detection of potential data set biases and can be used as a guide for a human expert, it does not allow an understanding of the underlying principles the model has learned. In this study, we instead show, to the best of our knowledge, for the first time results using feature visualization of neuroimaging CNNs. Particularly, we have trained CNNs for different tasks including sex classification and artificial lesion classification based on structural magnetic resonance imaging (MRI) data. We have then iteratively generated images that maximally activate specific neurons, in order to visualize the patterns they respond to. To improve the visualizations we compared several regularization strategies. The resulting images reveal the learned concepts of the artificial lesions, including their shapes, but remain hard to interpret for abstract features in the sex classification task.