论文标题
多任务学习的梯度手术
Gradient Surgery for Multi-Task Learning
论文作者
论文摘要
虽然深度学习和深度强化学习(RL)系统在图像分类,游戏玩法和机器人控制等领域表现出了令人印象深刻的结果,但数据效率仍然是一个主要挑战。多任务学习已成为一种有前途的方法,用于在多个任务中共享结构以实现更有效的学习。但是,多任务设置提出了许多优化挑战,因此与独立学习任务相比,很难实现巨大的效率提高。与单任务学习相比,多任务学习如此具有挑战性的原因尚未完全理解。在这项工作中,我们确定了导致有害梯度干扰的多任务优化领域的三个条件,并开发了一种简单但一般的方法来避免任务梯度之间的这种干扰。我们提出了一种梯度手术的形式,该形式将任务的梯度投射到任何其他具有冲突梯度的梯度的正常平面上。在一系列具有挑战性的多任务监督和多任务RL问题上,这种方法可带来效率和性能的巨大提高。此外,它是模型的不合SNOSTIC,可以与以前提供的多任务架构结合使用,以增强性能。
While deep learning and deep reinforcement learning (RL) systems have demonstrated impressive results in domains such as image classification, game playing, and robotic control, data efficiency remains a major challenge. Multi-task learning has emerged as a promising approach for sharing structure across multiple tasks to enable more efficient learning. However, the multi-task setting presents a number of optimization challenges, making it difficult to realize large efficiency gains compared to learning tasks independently. The reasons why multi-task learning is so challenging compared to single-task learning are not fully understood. In this work, we identify a set of three conditions of the multi-task optimization landscape that cause detrimental gradient interference, and develop a simple yet general approach for avoiding such interference between task gradients. We propose a form of gradient surgery that projects a task's gradient onto the normal plane of the gradient of any other task that has a conflicting gradient. On a series of challenging multi-task supervised and multi-task RL problems, this approach leads to substantial gains in efficiency and performance. Further, it is model-agnostic and can be combined with previously-proposed multi-task architectures for enhanced performance.