Self-supervised learning is a promising unsupervised learning framework that has achieved success with large floating point networks. But such networks are not readily deployable to edge devices. To accelerate deployment of models with the benefit of unsupervised representation learning to such resource limited devices for various downstream tasks, we propose a self-supervised learning method for binary networks that uses a moving target network. In particular, we propose to jointly train a randomly initialized classifier, attached to a pretrained floating point feature extractor, with a binary network. Additionally, we propose a feature similarity loss, a dynamic loss balancing and modified multi-stage training to further improve the accuracy, and call our method BURN. Our empirical validations over five downstream tasks using seven datasets show that BURN outperforms self-supervised baselines for binary networks and sometimes outperforms supervised pretraining.
翻译:自我监督的学习是一个充满希望的、不受监督的学习框架,它通过大型浮动点网络取得了成功。但这样的网络不能轻易被部署到边缘装置。为了加速部署模型,在不受监督的情况下向这种资源有限的下游任务设备学习各种下游任务,我们建议为使用移动目标网络的二进制网络采用自监督的学习方法。特别是,我们提议联合培训一个随机初始化的分类器,该分类器附属于一个预先训练的浮点特征提取器,并配有一个二进制网络。此外,我们提出了相似性损失、动态损失平衡和经过修改的多阶段培训,以进一步提高准确性,并调用我们的方法BURN。我们使用七套数据集对五个下游任务进行的经验验证表明,BURN比二进网络的自我监督基线更优,有时优于受监督的预培训。