Transfer learning can be seen as a data- and compute-efficient alternative to training models from scratch. The emergence of rich model repositories, such as TensorFlow Hub, enables practitioners and researchers to unleash the potential of these models across a wide range of downstream tasks. As these repositories keep growing exponentially, efficiently selecting a good model for the task at hand becomes paramount. By carefully comparing various selection and search strategies, we realize that no single method outperforms the others, and hybrid or mixed strategies can be beneficial. Therefore, we propose SHiFT, the first downstream task-aware, flexible, and efficient model search engine for transfer learning. These properties are enabled by a custom query language SHiFT-QL together with a cost-based decision maker, which we empirically validate. Motivated by the iterative nature of machine learning development, we further support efficient incremental executions of our queries, which requires a careful implementation when jointly used with our optimizations.
翻译:转移学习可以被视为从零开始就取代培训模式的数据和计算效率的替代方法。TensorFlow 枢纽等丰富的模型库的出现,使从业人员和研究人员能够在一系列广泛的下游任务中释放这些模型的潜力。随着这些储存库的成倍增长,高效地为手头的任务选择一个好的模型变得至关重要。通过仔细比较各种选择和搜索战略,我们认识到,没有任何一种方法比其他方法更优异,混合或混合战略是有好处的。因此,我们提议采用SHiFT,即第一个下游任务敏锐、灵活和高效的模型搜索引擎来进行转移学习。这些特性是由一种自定义查询语言SHiFT-QL以及一个基于成本的决策者一起促成的,我们通过实验来验证。受机器学习发展的迭接性质所激励,我们进一步支持高效地逐步执行我们的查询,这需要与优化一起使用时谨慎执行。