Robotic grasping of house-hold objects has made remarkable progress in recent years. Yet, human grasps are still difficult to synthesize realistically. There are several key reasons: (1) the human hand has many degrees of freedom (more than robotic manipulators); (2) the synthesized hand should conform to the surface of the object; and (3) it should interact with the object in a semantically and physically plausible manner. To make progress in this direction, we draw inspiration from the recent progress on learning-based implicit representations for 3D object reconstruction. Specifically, we propose an expressive representation for human grasp modelling that is efficient and easy to integrate with deep neural networks. Our insight is that every point in a three-dimensional space can be characterized by the signed distances to the surface of the hand and the object, respectively. Consequently, the hand, the object, and the contact area can be represented by implicit surfaces in a common space, in which the proximity between the hand and the object can be modelled explicitly. We name this 3D to 2D mapping as Grasping Field, parameterize it with a deep neural network, and learn it from data. We demonstrate that the proposed grasping field is an effective and expressive representation for human grasp generation. Specifically, our generative model is able to synthesize high-quality human grasps, given only on a 3D object point cloud. The extensive experiments demonstrate that our generative model compares favorably with a strong baseline and approaches the level of natural human grasps. Our method improves the physical plausibility of the hand-object contact reconstruction and achieves comparable performance for 3D hand reconstruction compared to state-of-the-art methods.
翻译:近些年来,对固定物体的机械掌握取得了显著的进展。然而,人类的掌握仍然难以现实地加以综合。有几个关键原因:(1) 人体手有多种程度的自由度(不仅仅是机器人操纵器);(2) 合成手应该与物体表面相符;(3) 合成手应该与物体的表面互动;(3) 它应该以静态和物理上合理的方式与物体发生互动;(3) 为了朝着这个方向取得进展,我们从最近基于学习的3D物体重建隐含表征的进展中汲取灵感。具体地说,我们提议对人体的掌握模型进行直观代表,这种模型既高效又容易与深层神经网络融合。我们的了解是,三维空间的每个点的每个点都可以以手与物体的表面距离为特征;因此,人工、物体和接触区域应该以隐含的表面为代表;我们提议的直径比方法只能以直径直的模型来显示我们人类的模型;我们提出的直径比方法只能以直截面的模型来显示我们人类的模型。