论文标题
视觉预测清晰的对象互动的先验
Visual Prediction of Priors for Articulated Object Interaction
论文作者
论文摘要
如果没有在类似领域的事先经验,新颖环境中的探索可能会具有挑战性。但是,人类能够快速有效地基于先前的经验。儿童在玩玩具时表现出这种行为。例如,给出一个带有黄色和蓝色门的玩具,一个孩子将没有明确的目标进行探索,但是一旦他们发现了如何打开黄门,他们很可能能够更快地打开蓝色的门。当进入厨房等新空间时,成年人也会表现出这种行为。我们开发了一种方法,上下文的先验预测,该预测提供了一种通过视觉在相似领域中的相互作用之间转移知识的方法。我们通过学习跨环境共享的视觉特征以及它们与行动的相关性来开发具有提高效率的探索性行为的代理。我们的问题被称为上下文中上下文是图像的上下文多臂强盗,并且机器人可以访问参数化的动作空间。鉴于一个新颖的对象,目的是通过几乎没有相互作用来最大化奖励。一个强烈表现出视觉特征和运动之间相关性的域是运动限制的机制。我们评估了模拟棱柱形和旋转关节的方法。
Exploration in novel settings can be challenging without prior experience in similar domains. However, humans are able to build on prior experience quickly and efficiently. Children exhibit this behavior when playing with toys. For example, given a toy with a yellow and blue door, a child will explore with no clear objective, but once they have discovered how to open the yellow door, they will most likely be able to open the blue door much faster. Adults also exhibit this behavior when entering new spaces such as kitchens. We develop a method, Contextual Prior Prediction, which provides a means of transferring knowledge between interactions in similar domains through vision. We develop agents that exhibit exploratory behavior with increasing efficiency, by learning visual features that are shared across environments, and how they correlate to actions. Our problem is formulated as a Contextual Multi-Armed Bandit where the contexts are images, and the robot has access to a parameterized action space. Given a novel object, the objective is to maximize reward with few interactions. A domain which strongly exhibits correlations between visual features and motion is kinemetically constrained mechanisms. We evaluate our method on simulated prismatic and revolute joints.