论文标题
学习最大独立集的推迟方法
Learning What to Defer for Maximum Independent Sets
论文作者
论文摘要
在各个科学领域,设计有效的组合优化算法出现。最近,深厚的加固学习(DRL)框架已成为一种新方法:它们可以自动化求解器的设计,同时较少依赖目标问题的复杂领域知识。但是,现有的DRL求解器使用与解决方案中元素数量成正比的许多阶段确定解决方案,这严重限制了它们对大规模图的适用性。在本文中,我们试图通过提出一种新颖的DRL方案,创造了推迟(LWD)来解决这个问题,在该方案中,代理人通过学习在每个阶段分发解决方案的元素决策来适应缩水或扩展阶段的数量。我们将提出的框架应用于最大独立集(MIS)问题,并证明了其对当前最新DRL方案的显着改善。我们还表明,在有限的时间预算下,LWD可以在具有数百万个顶点的大规模图上胜过传统的MIS求解器。
Designing efficient algorithms for combinatorial optimization appears ubiquitously in various scientific fields. Recently, deep reinforcement learning (DRL) frameworks have gained considerable attention as a new approach: they can automate the design of a solver while relying less on sophisticated domain knowledge of the target problem. However, the existing DRL solvers determine the solution using a number of stages proportional to the number of elements in the solution, which severely limits their applicability to large-scale graphs. In this paper, we seek to resolve this issue by proposing a novel DRL scheme, coined learning what to defer (LwD), where the agent adaptively shrinks or stretch the number of stages by learning to distribute the element-wise decisions of the solution at each stage. We apply the proposed framework to the maximum independent set (MIS) problem, and demonstrate its significant improvement over the current state-of-the-art DRL scheme. We also show that LwD can outperform the conventional MIS solvers on large-scale graphs having millions of vertices, under a limited time budget.