Upper limb and hand functionality is critical to many activities of daily living and the amputation of one can lead to significant functionality loss for individuals. From this perspective, advanced prosthetic hands of the future are anticipated to benefit from improved shared control between a robotic hand and its human user, but more importantly from the improved capability to infer human intent from multimodal sensor data to provide the robotic hand perception abilities regarding the operational context. Such multimodal sensor data may include various environment sensors including vision, as well as human physiology and behavior sensors including electromyography and inertial measurement units. A fusion methodology for environmental state and human intent estimation can combine these sources of evidence in order to help prosthetic hand motion planning and control. In this paper, we present a dataset of this type that was gathered with the anticipation of cameras being built into prosthetic hands, and computer vision methods will need to assess this hand-view visual evidence in order to estimate human intent. Specifically, paired images from human eye-view and hand-view of various objects placed at different orientations have been captured at the initial state of grasping trials, followed by paired video, EMG and IMU from the arm of the human during a grasp, lift, put-down, and retract style trial structure. For each trial, based on eye-view images of the scene showing the hand and object on a table, multiple humans were asked to sort in decreasing order of preference, five grasp types appropriate for the object in its given configuration relative to the hand. The potential utility of paired eye-view and hand-view images was illustrated by training a convolutional neural network to process hand-view images in order to predict eye-view labels assigned by humans.
翻译:上肢和手功能对于许多日常生活活动至关重要,一个人的截肢可以导致个人大量功能丧失。从这个角度出发,预期先进的假肢手未来会受益于机器人手和人体使用者之间更好的共同控制,但更重要的是,从多式传感器数据中推断人类意图的能力有所提高,以多式传感器数据为操作环境提供机器人手感知能力。这种多式传感器数据可能包括各种环境传感器,包括视觉,以及人体生理和行为传感器,包括电感学和惯性测量装置。环境状态和人类意图估计的混合方法可以结合这些证据来源,以帮助假肢手动规划和控制。在这个文件中,我们展示了这种类型的数据集,而这种数据集是随着照相机将建在假肢手掌上而收集的,计算机视觉方法需要评估这种手视视觉证据,以估计人类的操作环境。具体地,在最初的观察状态下,通过配对视频、EMG和IMU的图像,从人类眼睛臂部收集的数据集收集,在每张直径的图像中,在每张的图像上展示一个直径图,在每张的手图的图像中,在每张的手图的图的图的图,在手部中,在手部中显示中展示中,在手部中,在手部中,在手部中,在手部的手部中,在手部中,在手部的图上,在手部的图中,在手部的图中,在手部中,在手部显示一个直向下,在手部的图中,在手部的图中,在手部的图中,在手部的图上,在手部的图上,在手部的图中,在手部上,在手部,在手部,在手部上,在手部的图中,在手部,在手部,在手部,在手部,在手部,在手部,在手部,在手部的图中,在手部的图中,在手部的图上,在手部,在手部,在手部,在手部,在手部,在手部,在手部,在手部,在手部,在手部,在手部,在手部,在手部,在手部,在手部,在手