论文标题

合作环境和竞争环境的延迟意识到的多代理增强学习

Delay-Aware Multi-Agent Reinforcement Learning for Cooperative and Competitive Environments

论文作者

Chen, Baiming, Xu, Mengdi, Liu, Zuxin, Li, Liang, Zhao, Ding

论文摘要

行动和观察延迟普遍存在于现实世界中的网络物理系统中,这可能在增强学习设计中构成挑战。在处理多代理系统时,这尤其是一项艰巨的任务,其中一个代理的延迟可能会传播给其他代理。为了解决这个问题,本文提出了一个新颖的框架来处理延迟以及具有无模型深度强化学习的多代理任务的非平稳培训问题。我们正式定义了延迟感知的马尔可夫游戏,该游戏结合了环境中所有代理的延迟。为了解决延迟感知的马尔可夫游戏,我们应用集中式培训和分散执行,使代理商可以在培训期间使用额外的信息来缓解多机构系统的非平稳性问题,而无需执行期间集中的控制器。实验是在多代理粒子环境中进行的,包括合作通信,合作导航和竞争性实验。我们还测试了需要协调所有自动驾驶汽车以显示延迟意识的实际价值的交通情况下提出的算法。结果表明,拟议的延迟感知的多代理增强算法大大减轻了延迟引入的性能退化。代码和演示视频可在以下网址提供:https://github.com/baimingc/delay-aware-marl。

Action and observation delays exist prevalently in the real-world cyber-physical systems which may pose challenges in reinforcement learning design. It is particularly an arduous task when handling multi-agent systems where the delay of one agent could spread to other agents. To resolve this problem, this paper proposes a novel framework to deal with delays as well as the non-stationary training issue of multi-agent tasks with model-free deep reinforcement learning. We formally define the Delay-Aware Markov Game that incorporates the delays of all agents in the environment. To solve Delay-Aware Markov Games, we apply centralized training and decentralized execution that allows agents to use extra information to ease the non-stationarity issue of the multi-agent systems during training, without the need of a centralized controller during execution. Experiments are conducted in multi-agent particle environments including cooperative communication, cooperative navigation, and competitive experiments. We also test the proposed algorithm in traffic scenarios that require coordination of all autonomous vehicles to show the practical value of delay-awareness. Results show that the proposed delay-aware multi-agent reinforcement learning algorithm greatly alleviates the performance degradation introduced by delay. Codes and demo videos are available at: https://github.com/baimingc/delay-aware-MARL.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源