We study data-free knowledge distillation (KD) for monocular depth estimation (MDE), which learns a lightweight model for real-world depth perception tasks by compressing it from a trained teacher model while lacking training data in the target domain. Owing to the essential difference between image classification and dense regression, previous methods of data-free KD are not applicable to MDE. To strengthen its applicability in real-world tasks, in this paper, we propose to apply KD with out-of-distribution simulated images. The major challenges to be resolved are i) lacking prior information about object distribution of real-world training data, and ii) domain shift between simulated and real-world images. To cope with these difficulties, we propose a tailored framework for depth distillation. The framework generates new training samples for maximally covering distributed patterns of objects in the target domain and utilizes a transformation network to efficiently adapt them to the feature statistics preserved in the teacher model. Through extensive experiments on various depth estimation models and two different datasets, we show that our method outperforms the baseline KD by a good margin and even achieves slightly better performance with as few as 1/6 of training images, demonstrating a clear superiority.
翻译:我们研究无数据知识蒸馏法(KD),以进行单眼深度估计(MDE),这种蒸馏法从训练有素的教师模型中压缩,从缺乏目标领域培训数据的情况下,从缺乏培训的教师模型中压缩,从而为现实世界深度感知任务学习一种轻量级模型,从中学习一种轻量级模型,同时缺乏目标领域的培训数据模型;由于图像分类和密集回归之间的根本差异,以前没有数据的知识蒸馏法不适用于MDE。为了加强其在现实世界任务中的应用性,我们在本文件中建议应用这种方法,用模拟的模拟图像来加强其对真实世界性任务的适用性。需要解决的主要挑战是,一) 缺乏关于真实世界培训数据对象分布的事先信息,二) 模拟图像与现实世界图像之间的域位变化。为了应对这些困难,我们提出了一个专门设计的深度蒸馏框架。这个框架生成了新的培训样本,以尽量覆盖目标领域物体分布式样板,并利用一个改造网络来有效地使其适应教师模型中保存的特征统计。通过对各种深度估计模型和两个不同的数据集进行广泛的实验,我们所显示的方法比基准KD的距离要长得多,甚至比基准图像要稍好,并有明显的优势。