A neural network based flexible object manipulation system for a humanoid robot on FPGA is proposed. Although the manipulations of flexible objects using robots attract ever increasing attention since these tasks are the basic and essential activities in our daily life, it has been put into practice only recently with the help of deep neural networks. However such systems have relied on GPU accelerators, which cannot be implemented into the space limited robotic body. Although field programmable gate arrays (FPGAs) are known to be energy efficient and suitable for embedded systems, the model size should be drastically reduced since FPGAs have limited on-chip memory. To this end, we propose ``partially'' binarized deep convolutional auto-encoder technique, where only an encoder part is binarized to compress model size without degrading the inference accuracy. The model implemented on Xilinx ZCU102 achieves 41.1 frames per second with a power consumption of 3.1W, {\awano{which corresponds to 10x and 3.7x improvements from the systems implemented on Core i7 6700K and RTX 2080 Ti, respectively.
翻译:虽然使用机器人对灵活物体的操纵日益引起人们越来越多的注意,因为这些任务是我们日常生活中的基本和基本活动,但直到最近才在深神经网络的帮助下加以实践,然而,这些系统依赖GPU加速器,这些加速器无法在空间有限的机器人体中安装。虽然已知外地可编程门阵列节能且适合嵌入系统,但模型尺寸应大幅缩小,因为FPGA在芯片内存有限。为此,我们建议“部分地”将精密的二进式深共振自动电解器技术,其中只有一个元件在不降低导力精度的情况下被二进制成压缩模型大小。在 Xilinx ZCU102 上安装的模型每秒可安装41.1个框架,电耗量为3.1W, ~awano{, 相当于Core i7700K 和RTX 2080 Ti 上实施的系统的10x和3.7x改进。