论文标题

掌握领域:学习人类掌握的隐式表示

Grasping Field: Learning Implicit Representations for Human Grasps

论文作者

Karunratanakul, Korrawe, Yang, Jinlong, Zhang, Yan, Black, Michael, Muandet, Krikamol, Tang, Siyu

论文摘要

近年来,机器人抓住家用物体已取得了显着进展。然而,人类的掌握仍然很难现实地合成。有几个关键原因:(1)人类的手有许多自由度(不仅仅是机器人操纵者); (2)合成的手应符合物体表面; (3)它应该以语义和物理上合理的方式与对象相互作用。为了朝这个方向取得进展,我们从3D对象重建的基于学习的隐式表示方面的最新进展中汲取灵感。具体而言,我们为人类掌握建模提出了表达性表示,该表示有效且易于与深神经网络集成。我们的见解是,三维空间中的每个点都可以分别以签名的手和物体表面签名的距离为特征。因此,手,对象和接触区域可以用公共空间中隐式表面表示,在该空间中,可以将手与对象之间的接近性明确建模。我们将此3D至2D映射命名为“抓地”字段,用深层神经网络对其进行参数化,然后从数据中学习。我们证明,提出的抓地力领域是人类掌握产生的有效和表现力的表现。具体而言,我们的生成模型能够合成仅在3D对象点云上给出的高质量的人grasps。广泛的实验表明,我们的生成模型与强大的基线相比,并接近天然人类grASP的水平。与最新方法相比,我们的方法提高了手动触点重建的物理合理性,并在3D手重建方面实现了3D手重建的可比性。

Robotic grasping of house-hold objects has made remarkable progress in recent years. Yet, human grasps are still difficult to synthesize realistically. There are several key reasons: (1) the human hand has many degrees of freedom (more than robotic manipulators); (2) the synthesized hand should conform to the surface of the object; and (3) it should interact with the object in a semantically and physically plausible manner. To make progress in this direction, we draw inspiration from the recent progress on learning-based implicit representations for 3D object reconstruction. Specifically, we propose an expressive representation for human grasp modelling that is efficient and easy to integrate with deep neural networks. Our insight is that every point in a three-dimensional space can be characterized by the signed distances to the surface of the hand and the object, respectively. Consequently, the hand, the object, and the contact area can be represented by implicit surfaces in a common space, in which the proximity between the hand and the object can be modelled explicitly. We name this 3D to 2D mapping as Grasping Field, parameterize it with a deep neural network, and learn it from data. We demonstrate that the proposed grasping field is an effective and expressive representation for human grasp generation. Specifically, our generative model is able to synthesize high-quality human grasps, given only on a 3D object point cloud. The extensive experiments demonstrate that our generative model compares favorably with a strong baseline and approaches the level of natural human grasps. Our method improves the physical plausibility of the hand-object contact reconstruction and achieves comparable performance for 3D hand reconstruction compared to state-of-the-art methods.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源