论文标题
移动众包中的参与者选择问题的基于辅助任务的深度加固学习
Auxiliary-task Based Deep Reinforcement Learning for Participant Selection Problem in Mobile Crowdsourcing
论文作者
论文摘要
在移动众包(MCS)中,该平台选择参与者从旨在实现多个目标(例如,利润最大化,能源效率和公平性)的招聘人员那里完成位置感知的任务。但是,不同的MCS系统具有不同的目标,即使在一个MCS系统中,也可能存在冲突的目标。因此,设计适用于不同MCS系统以实现多个目标的参与者选择算法至关重要。为了解决这个问题,我们将参与者的选择问题作为增强学习问题,并提议用一种新颖的方法来解决它,我们称之为基于辅助任务的深度加强学习(ADRL)。我们使用变形金刚从MCS系统的上下文和指针网络中提取表示形式来处理组合优化问题。为了提高样本效率,我们采用了辅助任务培训过程,该过程训练网络以预测招聘人员的迫在眉睫的任务,从而有助于深度学习模型的学习。此外,我们在特定的MCS任务,乘车共享任务以及在此环境中进行广泛的绩效评估发布了模拟环境。实验结果表明,在各种环境中,ADRL优于其他公认的基线的样品效率优于和提高样品效率。
In mobile crowdsourcing (MCS), the platform selects participants to complete location-aware tasks from the recruiters aiming to achieve multiple goals (e.g., profit maximization, energy efficiency, and fairness). However, different MCS systems have different goals and there are possibly conflicting goals even in one MCS system. Therefore, it is crucial to design a participant selection algorithm that applies to different MCS systems to achieve multiple goals. To deal with this issue, we formulate the participant selection problem as a reinforcement learning problem and propose to solve it with a novel method, which we call auxiliary-task based deep reinforcement learning (ADRL). We use transformers to extract representations from the context of the MCS system and a pointer network to deal with the combinatorial optimization problem. To improve the sample efficiency, we adopt an auxiliary-task training process that trains the network to predict the imminent tasks from the recruiters, which facilitates the embedding learning of the deep learning model. Additionally, we release a simulated environment on a specific MCS task, the ride-sharing task, and conduct extensive performance evaluations in this environment. The experimental results demonstrate that ADRL outperforms and improves sample efficiency over other well-recognized baselines in various settings.