论文标题

专家问题的记忆范围

Memory Bounds for the Experts Problem

论文作者

Srinivas, Vaidehi, Woodruff, David P., Xu, Ziyu, Zhou, Samson

论文摘要

通过专家建议在线学习是顺序预测的基本问题。在此问题中,该算法可以访问每天进行预测的$ n $“专家”。每天的目标是处理这些预测,并以最低成本进行预测。在进行预测之后,该算法在当天看到了实际结果,更新了其状态,然后转到第二天。通过与该集合中最好的专家相比,它的表现如何,判断出一种算法。 此问题的经典算法是乘法算法。但是,据我们所知,每个应用程序都依赖于为每个专家存储权重,并使用$ω(n)$内存。在自然流媒体模型中,了解使用专家建议问题或运行标准顺序预测算法来解决在线学习所需的记忆的工作很少,当专家的数量以及专家做出预测的天数时,这一点尤其重要。 我们在流式设置中使用专家建议问题开始学习学习,并显示下限和上限。我们用于I.I.D.的下限,随机顺序和对抗顺序流使用新颖的掩蔽技术对定制问题进行了简化,以表现出平稳的折衷,以使后悔与记忆。我们的上限显示了在专家的小“池”上运行标准顺序预测算法的新颖方法,从而减少了必要的内存。对于随机顺序,我们表明我们的上限紧密到低订单项。我们希望这些结果和技术将在在线学习中具有广泛的应用程序,并且可以基于标准顺序预测技术(如乘法权重)激发算法,以解决内存约束设置中的许多其他问题。

Online learning with expert advice is a fundamental problem of sequential prediction. In this problem, the algorithm has access to a set of $n$ "experts" who make predictions on each day. The goal on each day is to process these predictions, and make a prediction with the minimum cost. After making a prediction, the algorithm sees the actual outcome on that day, updates its state, and then moves on to the next day. An algorithm is judged by how well it does compared to the best expert in the set. The classical algorithm for this problem is the multiplicative weights algorithm. However, every application, to our knowledge, relies on storing weights for every expert, and uses $Ω(n)$ memory. There is little work on understanding the memory required to solve the online learning with expert advice problem, or run standard sequential prediction algorithms, in natural streaming models, which is especially important when the number of experts, as well as the number of days on which the experts make predictions, is large. We initiate the study of the learning with expert advice problem in the streaming setting, and show lower and upper bounds. Our lower bound for i.i.d., random order, and adversarial order streams uses a reduction to a custom-built problem using a novel masking technique, to show a smooth trade-off for regret versus memory. Our upper bounds show novel ways to run standard sequential prediction algorithms in rounds on small "pools" of experts, thus reducing the necessary memory. For random-order streams, we show that our upper bound is tight up to low order terms. We hope that these results and techniques will have broad applications in online learning, and can inspire algorithms based on standard sequential prediction techniques, like multiplicative weights, for a wide range of other problems in the memory-constrained setting.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源