The number of mobile robots with constrained computing resources that need to execute complex machine learning models has been increasing during the past decade. Commonly, these robots rely on edge infrastructure accessible over wireless communication to execute heavy computational complex tasks. However, the edge might become unavailable and, consequently, oblige the execution of the tasks on the robot. This work focuses on making it possible to execute the tasks on the robots by reducing the complexity and the total number of parameters of pre-trained computer vision models. This is achieved by using model compression techniques such as Pruning and Knowledge Distillation. These compression techniques have strong theoretical and practical foundations, but their combined usage has not been widely explored in the literature. Therefore, this work especially focuses on investigating the effects of combining these two compression techniques. The results of this work reveal that up to 90% of the total number of parameters of a computer vision model can be removed without any considerable reduction in the model's accuracy.
翻译:在过去十年中,需要实施复杂机器学习模型的计算机资源有限的移动机器人数量一直在增加。通常,这些机器人依靠通过无线通信获得的边缘基础设施来执行繁重的计算复杂任务。然而,边缘可能变得无法使用,因此,必须执行机器人的任务。这项工作的重点是通过降低预先培训的计算机视觉模型的复杂性和参数总数,使机器人能够执行任务。这是通过使用模型压缩技术(如普林和知识蒸馏)实现的。这些压缩技术具有很强的理论和实践基础,但文献中并未广泛探讨它们的结合使用。因此,这项工作特别侧重于调查这两种压缩技术的结合效果。这项工作的结果显示,在不大幅降低模型准确性的情况下,计算机视觉模型总参数的90%都可以被删除。