论文标题
除了解释:基于XAI的模型改进的机遇和挑战
Beyond Explaining: Opportunities and Challenges of XAI-Based Model Improvement
论文作者
论文摘要
可解释的人工智能(XAI)是一个新兴的研究领域,为高度复杂而不透明的机器学习(ML)模型带来了透明度。尽管开发了多种方法来解释近年来黑盒分类器的决策,但这些工具很少超出可视化目的。直到最近,研究人员才开始在实践中采用解释来实际改善模型。本文提供了有关XAI实际上用于改善ML模型的各种属性的技术的全面概述,并系统地对这些方法进行了分类,并比较了它们各自的优势和劣势。我们提供了有关这些方法的理论观点,并通过对玩具和现实环境的实验进行经验来展示解释如何帮助改善诸如模型概括能力或推理等属性等属性。我们进一步讨论了这些方法的潜在警告和缺点。我们得出的结论是,尽管基于XAI的模型改进即使对复杂且不容易量化的模型属性也具有显着的有益效果,但需要仔细应用这些方法,因为它们的成功可能会根据多种因素而变化,例如所使用的模型和数据集,或使用的解释方法。
Explainable Artificial Intelligence (XAI) is an emerging research field bringing transparency to highly complex and opaque machine learning (ML) models. Despite the development of a multitude of methods to explain the decisions of black-box classifiers in recent years, these tools are seldomly used beyond visualization purposes. Only recently, researchers have started to employ explanations in practice to actually improve models. This paper offers a comprehensive overview over techniques that apply XAI practically for improving various properties of ML models, and systematically categorizes these approaches, comparing their respective strengths and weaknesses. We provide a theoretical perspective on these methods, and show empirically through experiments on toy and realistic settings how explanations can help improve properties such as model generalization ability or reasoning, among others. We further discuss potential caveats and drawbacks of these methods. We conclude that while model improvement based on XAI can have significant beneficial effects even on complex and not easily quantifyable model properties, these methods need to be applied carefully, since their success can vary depending on a multitude of factors, such as the model and dataset used, or the employed explanation method.