论文标题
对自动驾驶汽车运动计划的深入强化学习调查
Survey of Deep Reinforcement Learning for Motion Planning of Autonomous Vehicles
论文作者
论文摘要
近年来,自动驾驶汽车领域的学术研究已广受欢迎,与传感器技术,V2X通信,安全,安全,决策,控制,甚至法律和标准化规则有关。除了经典的控制设计方法外,几乎所有这些领域都存在人工智能和机器学习方法。研究的另一部分侧重于运动计划的不同层面,例如战略决策,轨迹计划和控制。已经开发了机器学习中广泛的技术,本文描述了这些领域之一,深度强化学习(DRL)。本文提供了对分层运动计划问题的见解,并描述了DRL的基础知识。设计这样的系统的主要要素是环境的建模,建模抽象,国家的描述和感知模型,适当的奖励以及基础神经网络的实现。本文描述了车辆模型,模拟可能性和计算要求。介绍了对不同层和观察模型的战略决策,例如连续和离散的状态表示,基于网格和基于摄像机的解决方案。该论文通过自动驾驶的不同任务和级别(例如,遵循汽车的驾驶,保持车道,轨迹,关注,合并或驾驶繁殖量)进行了系统化的最先进解决方案。最后,讨论了开放的问题和未来挑战。
Academic research in the field of autonomous vehicles has reached high popularity in recent years related to several topics as sensor technologies, V2X communications, safety, security, decision making, control, and even legal and standardization rules. Besides classic control design approaches, Artificial Intelligence and Machine Learning methods are present in almost all of these fields. Another part of research focuses on different layers of Motion Planning, such as strategic decisions, trajectory planning, and control. A wide range of techniques in Machine Learning itself have been developed, and this article describes one of these fields, Deep Reinforcement Learning (DRL). The paper provides insight into the hierarchical motion planning problem and describes the basics of DRL. The main elements of designing such a system are the modeling of the environment, the modeling abstractions, the description of the state and the perception models, the appropriate rewarding, and the realization of the underlying neural network. The paper describes vehicle models, simulation possibilities and computational requirements. Strategic decisions on different layers and the observation models, e.g., continuous and discrete state representations, grid-based, and camera-based solutions are presented. The paper surveys the state-of-art solutions systematized by the different tasks and levels of autonomous driving, such as car-following, lane-keeping, trajectory following, merging, or driving in dense traffic. Finally, open questions and future challenges are discussed.