We propose a real-time vision-based teleoperation approach for robotic arms that employs a single depth-based camera, exempting the user from the need for any wearable devices. By employing a natural user interface, this novel approach leverages the conventional fine-tuning control, turning it into a direct body pose capture process. The proposed approach is comprised of two main parts. The first is a nonlinear customizable pose mapping based on Thin-Plate Splines (TPS), to directly transfer human body motion to robotic arm motion in a nonlinear fashion, thus allowing matching dissimilar bodies with different workspace shapes and kinematic constraints. The second is a Deep Neural Network hand-state classifier based on Long-term Recurrent Convolutional Networks (LRCN) that exploits the temporal coherence of the acquired depth data. We validate, evaluate and compare our approach through both classical cross-validation experiments of the proposed hand state classifier; and user studies over a set of practical experiments involving variants of pick-and-place and manufacturing tasks. Results revealed that LRCN networks outperform single image Convolutional Neural Networks; and that users' learning curves were steep, thus allowing the successful completion of the proposed tasks. When compared to a previous approach, the TPS approach revealed no increase in task complexity and similar times of completion, while providing more precise operation in regions closer to workspace boundaries.
翻译:我们建议对机器人武器采用实时的基于视觉的远程操作方法,该方法使用单一的深底相机,使用户免于需要任何磨损装置。 通过使用自然用户界面,这种新颖的方法利用常规微调控制,将它变成直接体形捕捉过程。拟议方法由两个主要部分组成。第一个是非线性定制的基于Thin-Plate Splines(TPS)的图像绘图,以非线性方式将人体运动直接转移到机器人臂运动,从而允许将不同机构与不同工作空间形状和动态限制相匹配。第二个是深神经网络手状态分类器,以长期持续革命网络为基础,利用获得的深度数据的时间一致性。我们通过对拟议手级分类(TPS)进行经典的交叉校验试验来验证、评价和比较我们的方法;对一系列实际实验的用户研究,涉及选址和制造任务的变异体,从而可以将不同机构与不同的工作形状和动态限制相匹配。第二个是深层网络的深层神经网络(深层网络)的手划分级分类,利用获得的时,从而将以往完成任务的成功时间进行对比。