The images captured by vision-based tactile sensors carry information about high-resolution tactile fields, such as the distribution of the contact forces applied to their soft sensing surface. However, extracting the information encoded in the images is challenging and often addressed with learning-based approaches, which generally require a large amount of training data. This article proposes a strategy to generate tactile images in simulation for a vision-based tactile sensor based on an internal camera that tracks the motion of spherical particles within a soft material. The deformation of the material is simulated in a finite element environment under a diverse set of contact conditions, and spherical particles are projected to a simulated image. Features extracted from the images are mapped to the 3D contact force distribution, with the ground truth also obtained via finite-element simulations, with an artificial neural network that is therefore entirely trained on synthetic data avoiding the need for real-world data collection. The resulting model exhibits high accuracy when evaluated on real-world tactile images, is transferable across multiple tactile sensors without further training, and is suitable for efficient real-time inference.
翻译:以视觉为基础的触动传感器所拍摄的图像含有高分辨率触动场的信息,例如用于软感应表面的接触力量的分布。然而,提取图像中所编码的信息具有挑战性,而且往往以学习为基础的方法加以解决,通常需要大量的培训数据。本文章提出了一个战略,在模拟基于视觉的触动传感器时生成触动图像,以模拟基于视觉的触动感应传感器,以一个内部相机为基础,跟踪软材料中球状粒的移动。材料的变形在一套不同的接触条件下的有限元素环境中进行模拟,并且将球状颗粒投向一个模拟图像。从图像中提取的特征被映射到3D接触力分布图中,通过限定元素模拟也获得了地面真相,因此,人工神经网络经过全面培训,无需收集真实世界数据。所产生的模型在对真实世界触动图像进行评估时显示高度准确性,不经进一步培训就可跨越多个触动感应传感器,适合高效的实时反射。