Neural fields can be trained from scratch to represent the shape and appearance of 3D scenes efficiently. It has also been shown that they can densely map correlated properties such as semantics, via sparse interactions from a human labeller. In this work, we show that a robot can densely annotate a scene with arbitrary discrete or continuous physical properties via its own fully-autonomous experimental interactions, as it simultaneously scans and maps it with an RGB-D camera. A variety of scene interactions are possible, including poking with force sensing to determine rigidity, measuring local material type with single-pixel spectroscopy or predicting force distributions by pushing. Sparse experimental interactions are guided by entropy to enable high efficiency, with tabletop scene properties densely mapped from scratch in a few minutes from a few tens of interactions.
翻译:从头到尾可以训练神经场域,以高效地代表 3D 场景的形状和外观。 也已经表明它们可以通过人体标签员的微薄互动, 密集地映射诸如语义学等相关属性。 在这项工作中, 我们显示机器人可以通过其自身的完全自主的实验互动, 同时扫描和用 RGB- D 相机绘制场景图。 各种场景互动是可能的, 包括用强力感测来确定僵硬性, 用单像素光谱测量本地物质类型, 或用推力预测力量分布。 粗微的实验互动以昆虫为指南, 以便实现高效, 桌面场景特性从零到零的几分钟内从几数十次互动中集中绘制出来。