论文标题

一项关于可解释的强化学习的调查:概念,算法,挑战

A Survey on Explainable Reinforcement Learning: Concepts, Algorithms, Challenges

论文作者

Qing, Yunpeng, Liu, Shunyu, Song, Jie, Wang, Huiqiong, Song, Mingli

论文摘要

强化学习(RL)是一种流行的机器学习范式,智能代理与环境互动以实现长期目标。在深度学习的复兴驱动下,Deep RL(DRL)在各种复杂的控制任务中都取得了巨大的成功。尽管取得了令人鼓舞的结果,但深厚的基于神经网络的主链被广泛认为是一个黑匣子,它阻碍了从业者在高安全性和可靠性至关重要的现实情况下信任和雇用训练有素的代理。为了减轻这个问题,已经提出了大量的文献来阐明智能代理的内部运作,通过构建内在的可解释性或事后解释性。在这项调查中,我们对可解释的RL(XRL)的现有作品进行了全面的审查,并引入了新的分类法,其中将先前的作品明确分为模型解释,奖励解释,国家解释和任务解释方法。我们还审查并强调了RL方法,这些方法相反,它利用人类知识来促进代理的学习效率和表现,而这种方法在XRL领域经常被忽略。讨论了XRL中的一些挑战和机遇。这项调查旨在提供XRL的高级摘要,并激发对更有效XRL解决方案的未来研究。在https://github.com/plankson/awesome-explainable-reinable-forecement-learning中收集并分类相应的开源代码。

Reinforcement Learning (RL) is a popular machine learning paradigm where intelligent agents interact with the environment to fulfill a long-term goal. Driven by the resurgence of deep learning, Deep RL (DRL) has witnessed great success over a wide spectrum of complex control tasks. Despite the encouraging results achieved, the deep neural network-based backbone is widely deemed as a black box that impedes practitioners to trust and employ trained agents in realistic scenarios where high security and reliability are essential. To alleviate this issue, a large volume of literature devoted to shedding light on the inner workings of the intelligent agents has been proposed, by constructing intrinsic interpretability or post-hoc explainability. In this survey, we provide a comprehensive review of existing works on eXplainable RL (XRL) and introduce a new taxonomy where prior works are clearly categorized into model-explaining, reward-explaining, state-explaining, and task-explaining methods. We also review and highlight RL methods that conversely leverage human knowledge to promote learning efficiency and performance of agents while this kind of method is often ignored in XRL field. Some challenges and opportunities in XRL are discussed. This survey intends to provide a high-level summarization of XRL and to motivate future research on more effective XRL solutions. Corresponding open source codes are collected and categorized at https://github.com/Plankson/awesome-explainable-reinforcement-learning.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源