论文标题
基于信任的多代理增强学习系统的共识
Trust-based Consensus in Multi-Agent Reinforcement Learning Systems
论文作者
论文摘要
多代理增强学习(MARL)中经常被忽视的问题是环境中不可靠的代理的潜在存在,其与预期行为的偏差可以阻止系统完成其预期的任务。特别是,共识是合作分布式多代理系统的基本基础问题。共识需要位于分散的通信网络中的不同代理,才能从他们提出的一系列初始建议中达成协议。基于学习的代理人应采用一项协议,尽管系统中有一个或多个不可靠的代理,但仍可以达成共识。本文考虑了共识是一个案例研究,研究了MARL中不可靠的药物的问题。与分布式系统文献中的既定结果相呼应,我们的实验表明,即使是适度的这种代理也可以极大地影响在网络环境中达成共识的能力。我们提出了基于加强学习的信任共识(RLTC),这是一种分散的信任机制,在该机制中,代理可以独立地决定与哪些邻居进行交流。我们从经验上证明,我们的信任机制能够有效地处理不可靠的代理,这是由更高的共识成功率证明的。
An often neglected issue in multi-agent reinforcement learning (MARL) is the potential presence of unreliable agents in the environment whose deviations from expected behavior can prevent a system from accomplishing its intended tasks. In particular, consensus is a fundamental underpinning problem of cooperative distributed multi-agent systems. Consensus requires different agents, situated in a decentralized communication network, to reach an agreement out of a set of initial proposals that they put forward. Learning-based agents should adopt a protocol that allows them to reach consensus despite having one or more unreliable agents in the system. This paper investigates the problem of unreliable agents in MARL, considering consensus as a case study. Echoing established results in the distributed systems literature, our experiments show that even a moderate fraction of such agents can greatly impact the ability of reaching consensus in a networked environment. We propose Reinforcement Learning-based Trusted Consensus (RLTC), a decentralized trust mechanism, in which agents can independently decide which neighbors to communicate with. We empirically demonstrate that our trust mechanism is able to handle unreliable agents effectively, as evidenced by higher consensus success rates.