In spite of the great success of deep learning technologies, training and delivery of a practically serviceable model is still a highly time-consuming process. Furthermore, a resulting model is usually too generic and heavyweight, and hence essentially goes through another expensive model compression phase to fit in a resource-limited device like embedded systems. Inspired by the fact that a machine learning task specifically requested by mobile users is often much simpler than it is supported by a massive generic model, this paper proposes a framework, called Pool of Experts (PoE), that instantly builds a lightweight and task-specific model without any training process. For a realtime model querying service, PoE first extracts a pool of primitive components, called experts, from a well-trained and sufficiently generic network by exploiting a novel conditional knowledge distillation method, and then performs our train-free knowledge consolidation to quickly combine necessary experts into a lightweight network for a target task. Thanks to this train-free property, in our thorough empirical study, PoE can build a fairly accurate yet compact model in a realtime manner, whereas it takes a few minutes per query for the other training methods to achieve a similar level of the accuracy.
翻译:尽管深层学习技术取得了巨大成功,但培训和提供实用模型仍是一个非常耗时的过程。此外,所产生的模型通常过于通用和重量重,因此基本上要经历另一个昂贵的模型压缩阶段,以适合像嵌入系统这样的资源有限的设备。由于移动用户特别要求的机器学习任务往往比大规模通用模型要简单得多,本文件提议了一个称为“专家人才库(PoE)”的框架,这个框架可以不经过任何培训程序即刻建立一个轻巧和具体任务模型。对于实时模型查询服务,PoE首先从一个经过良好训练的、足够通用的网络中抽出一批原始组件,即专家,利用一种新的有条件的知识蒸馏方法,然后进行我们的无培训知识整合,以便迅速将必要的专家纳入一个轻量网络,完成一个目标任务。由于这个没有培训的产业,PoE可以在我们彻底的经验研究中实时地建立一个相当准确但又紧凑的模型,而其他培训方法则需要几分钟的询问,以便达到类似的精确程度。