论文标题
基于梯度的显着图在深度加强学习中是否有用?
Are Gradient-based Saliency Maps Useful in Deep Reinforcement Learning?
论文作者
论文摘要
深度增强学习(DRL)将经典的增强学习算法与深度神经网络联系起来。 DRL中的一个问题是CNN是黑盒,很难理解代理的决策过程。为了能够在高度危险的环境中为人类和机器使用RL代理,开发人员需要一个调试工具来确保代理商执行预期。当前,奖励主要用于解释代理商的学习能力。但是,如果代理人通过记住政策而不是学习对环境做出回应,这可能会导致欺骗性结论。在这项工作中,可以证明可以借助梯度可视化技术来识别此问题。这项工作将图像分类领域的一些最著名的可视化方法带到了深度强化学习领域。此外,已经开发了两种新的可视化技术,其中一种提供了特别好的结果。它被证明在何种程度上可以在增强学习领域使用。同样,问题是关于通过不同可视化技术在不同环境中可视化的DRL算法的情况。
Deep Reinforcement Learning (DRL) connects the classic Reinforcement Learning algorithms with Deep Neural Networks. A problem in DRL is that CNNs are black-boxes and it is hard to understand the decision-making process of agents. In order to be able to use RL agents in highly dangerous environments for humans and machines, the developer needs a debugging tool to assure that the agent does what is expected. Currently, rewards are primarily used to interpret how well an agent is learning. However, this can lead to deceptive conclusions if the agent receives more rewards by memorizing a policy and not learning to respond to the environment. In this work, it is shown that this problem can be recognized with the help of gradient visualization techniques. This work brings some of the best-known visualization methods from the field of image classification to the area of Deep Reinforcement Learning. Furthermore, two new visualization techniques have been developed, one of which provides particularly good results. It is being proven to what extent the algorithms can be used in the area of Reinforcement learning. Also, the question arises on how well the DRL algorithms can be visualized across different environments with varying visualization techniques.