论文标题
对深泰勒分解的严格研究
A Rigorous Study Of The Deep Taylor Decomposition
论文作者
论文摘要
显着性方法试图通过突出样本的最显着特征来解释深层神经网络。一些广泛使用的方法基于一个称为深泰勒分解(DTD)的理论框架,该框架将泰勒定理的递归应用正式应用于网络层。但是,最近的工作发现这些方法独立于网络的更深层,并且似乎仅对较低级别的图像结构做出反应。在这里,我们研究了DTD理论,以更好地理解这种困惑的行为,并发现当Taylor root Point(用户选择的算法的重要参数)局部恒定时,深层分解与基本梯度$ \ times $输入方法相当。如果根点是局部输入依赖性的,则可以证明任何解释是合理的。在这种情况下,该理论不受限制。在经验评估中,我们发现DTD根与输入相同的线性区域不存在 - 与泰勒定理的基本假设相反。将DTD的理论基础作为解释的可靠性来源。但是,我们的发现敦促提出此类主张。
Saliency methods attempt to explain deep neural networks by highlighting the most salient features of a sample. Some widely used methods are based on a theoretical framework called Deep Taylor Decomposition (DTD), which formalizes the recursive application of the Taylor Theorem to the network's layers. However, recent work has found these methods to be independent of the network's deeper layers and appear to respond only to lower-level image structure. Here, we investigate the DTD theory to better understand this perplexing behavior and found that the Deep Taylor Decomposition is equivalent to the basic gradient$\times$input method when the Taylor root points (an important parameter of the algorithm chosen by the user) are locally constant. If the root points are locally input-dependent, then one can justify any explanation. In this case, the theory is under-constrained. In an empirical evaluation, we find that DTD roots do not lie in the same linear regions as the input - contrary to a fundamental assumption of the Taylor theorem. The theoretical foundations of DTD were cited as a source of reliability for the explanations. However, our findings urge caution in making such claims.